IBM RED BOOKS SERIES SAN's
IBM RED BOOKS SERIES SAN's
Introduction to
Storage Area Networks
Jon Tate
Pall Beck
Hector Hugo Ibarra
Shanmuganathan Kumaravel
Libor Miklas
Redbooks
SG24-5470-06
Note: Before using this information and the product it supports, read the information in Notices on
page ix.
Copyright International Business Machines Corporation 1999, 2016. All rights reserved.
Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule
Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .x
Summary of changes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
January 2016, Seventh Edition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
IBM Redbooks promotions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Chapter 1. Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.1 The importance of communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Interconnection models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2.1 The open systems interconnection model. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2.2 Translating the OSI model to the physical world. . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3 Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.1 Storing data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.2 Redundant Array of Independent Disks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.4 Storage area networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.5 Storage area network components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.5.1 Storage area network connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.5.2 Storage area network storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.5.3 Storage area network servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.6 The importance of standards or models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
Chapter 2. Storage area networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.1 Storage area networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.1.1 The problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.1.2 Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2 Using a storage area network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2.1 Infrastructure simplification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2.2 Information lifecycle management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2.3 Business continuity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3 Using the storage area network components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.1 Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.2 Storage area network connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.3 Servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.4 Putting the components together . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
17
18
18
19
20
20
21
22
22
22
23
28
32
33
34
34
35
38
iii
iv
3.2 Layers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3 Optical cables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.1 Attenuation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.2 Maximum power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.3 Fiber in the storage area network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.3.4 Dark fiber . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4 Classes of service . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.1 Class 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.2 Class 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.3 Class 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.4 Class 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.5 Class 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.6 Class 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.4.7 Class F . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.5 Fibre Channel data movement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.5.1 Byte-encoding schemes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6 Data transport . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.1 Ordered set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.2 Frames . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.3 Sequences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.4 Exchanges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.5 In order and out of order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.6 Latency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.6.7 Open fiber control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.7 Flow control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.7.1 Buffer to buffer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.7.2 End to end. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.7.3 Controlling the flow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
3.7.4 Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
40
43
43
43
44
49
49
50
50
50
50
50
51
51
51
52
54
54
55
57
57
58
59
59
60
60
60
60
60
63
64
64
65
65
67
68
68
71
73
75
75
78
78
80
80
80
80
80
81
81
81
81
82
83
85
85
85
86
87
87
88
88
Contents
vi
129
130
131
131
132
134
135
135
138
139
139
141
142
144
144
145
146
149
149
149
151
152
153
154
154
156
156
157
157
157
157
158
158
158
159
159
163
165
167
168
168
168
170
171
171
171
172
Chapter 8. Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
8.1 Management principles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
8.1.1 Management types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
8.1.2 Connecting to storage area network management tools. . . . . . . . . . . . . . . . . . .
8.1.3 Storage area network fault isolation and troubleshooting . . . . . . . . . . . . . . . . . .
173
174
174
176
176
177
177
179
180
180
182
183
183
184
185
185
Chapter 9. Security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.1 Security in the storage area network. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2 Security principles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2.1 Access control. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2.2 Auditing and accounting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2.3 Data security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2.4 Securing a fabric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.2.5 Zoning, masking, and binding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.3 Data security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4 Storage area network encryption . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.1 Basic encryption definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.2 Data-in-flight . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.3 Data-at-rest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.4 Digital certificates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.5 Encryption algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.6 Key management considerations and security standards. . . . . . . . . . . . . . . . . .
9.4.7 b-type encryption methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.4.8 Cisco encryption methods. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.5 Encryption standards and algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9.6 Security common practices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
193
194
195
195
195
195
196
197
198
198
198
200
201
201
201
202
203
205
207
208
209
210
210
210
211
211
214
215
216
216
219
221
225
225
227
228
229
230
230
232
Contents
vii
viii
235
236
237
238
239
239
240
241
243
243
245
Chapter 12. IBM Fibre Channel storage area network product portfolio . . . . . . . . . .
12.1 Classification of IBM SAN products . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.2 SAN Fibre Channel networking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.3 Entry SAN switches. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.3.1 IBM System Storage SAN24B-5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.3.2 IBM System Storage SAN24B-4 Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.4 Midrange SAN switches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.4.1 Cisco MDS 9396S 16G Multilayer Fabric Switch . . . . . . . . . . . . . . . . . . . . . . .
12.4.2 IBM System Networking SAN96B-5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.4.3 IBM System Storage SAN48B-5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.4.4 Cisco MDS 9148S 16G Multilayer Fabric Switch for IBM System Storage. . . .
12.5 Enterprise SAN directors. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.5.1 Cisco MDS 9710 Multilayer Director . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.5.2 IBM System Storage SAN384B-2 and SAN768B-2 . . . . . . . . . . . . . . . . . . . . .
12.5.3 Cisco MDS 9706 Multilayer Director for IBM System Storage . . . . . . . . . . . . .
12.5.4 Cisco MDS 9500 Series Multilayer Directors . . . . . . . . . . . . . . . . . . . . . . . . . .
12.6 Specialty switches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.6.1 IBM System Storage SAN42B-R. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12.6.2 Cisco MDS 9250i Multiservice Fabric Switch . . . . . . . . . . . . . . . . . . . . . . . . . .
12.6.3 IBM System Storage SAN06B-R. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
247
248
248
249
249
250
251
251
253
254
255
256
256
258
262
263
266
266
267
268
271
272
272
272
273
273
273
274
274
274
Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
275
275
276
277
Notices
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the users responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION AS IS
WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED
TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A
PARTICULAR PURPOSE. Some jurisdictions do not allow disclaimer of express or implied warranties in
certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBMs future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided AS IS, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
Copyright IBM Corp. 1999, 2016. All rights reserved.
ix
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines
Corporation, registered in many jurisdictions worldwide. Other product and service names might be
trademarks of IBM or other companies. A current list of IBM trademarks is available on the web at Copyright
and trademark information at http://www.ibm.com/legal/copytrade.shtml
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
AFS
AIX
DB2
Domino
DS8000
Easy Tier
ECKD
FICON
GPFS
HACMP
HyperFactor
IBM
IBM z
IBM z Systems
Informix
Lotus
OS/390
Power Systems
PowerHA
ProtecTIER
Redbooks
Redbooks (logo)
Storwize
System Storage
System z
System z9
Tivoli
z Systems
z/OS
z9
Summary of changes
This section describes the technical changes made in this edition of the book and in previous
editions. This edition might also include minor corrections and editorial changes that are not
identified.
Summary of Changes
for SG24-5470-06
for Introduction to Storage Area Networks
as created or updated on January 14, 2016.
New information
Chapter 12 contains the current IBM SAN portfolio at time of updating this book
(November). Editorial changes and corrections are included.
Changed information
Products that were removed from the IBM portfolio are deleted from the book.
xi
xii
Download
Now
Android
iOS
ibm.com/Redbooks
About Redbooks
Preface
The superabundance of data that is created by todays businesses is making storage a
strategic investment priority for companies of all sizes. As storage takes precedence, the
following major initiatives emerge:
Flatten and converge your network: IBM takes an open, standards-based approach to
implement the latest advances in the flat, converged data center network designs of today.
IBM Storage solutions enable clients to deploy a high-speed, low-latency Unified Fabric
Architecture.
Optimize and automate virtualization: Advanced virtualization awareness reduces the cost
and complexity of deploying physical and virtual data center infrastructure.
Simplify management: IBM data center networks are easy to deploy, maintain, scale, and
virtualize, delivering the foundation of consolidated operations for dynamic infrastructure
management.
Storage is no longer an afterthought. Too much is at stake. Companies are searching for
more ways to efficiently manage expanding volumes of data, and to make that data
accessible throughout the enterprise. This demand is propelling the move of storage into the
network. Also, the increasing complexity of managing large numbers of storage devices and
vast amounts of data is driving greater business value into software and services.
With current estimates of the amount of data to be managed and made available increasing at
60% each year, this outlook is where a storage area network (SAN) enters the arena. SANs
are the leading storage infrastructure for the global economy of today. SANs offer simplified
storage management, scalability, flexibility, and availability; and improved data access,
movement, and backup.
Welcome to the cognitive era.
The smarter data center with the improved economics of IT can be achieved by connecting
servers and storage with a high-speed and intelligent network fabric. A smarter data center
that hosts IBM Storage solutions can provide an environment that is smarter, faster, greener,
open, and easy to manage.
This IBM Redbooks publication provides an introduction to SAN and Ethernet networking,
and how these networks help to achieve a smarter data center. This book is intended for
people who are not very familiar with IT, or who are just starting out in the IT world.
Also, be sure to see the IBM Storage Redbooks portal for the latest material from the
International Technical Support Organization (ITSO):
http://www.redbooks.ibm.com/portals/storage
xv
Authors
This book was produced by a team of specialists from around the world working at the
International Technical Support Organization, Poughkeepsie Center.
Jon Tate is a Project Manager for IBM System Storage SAN
Solutions at the International Technical Support Organization
(ITSO), San Jose Center. Before Jon joined the ITSO in 1999, he
worked in the IBM Technical Support Center, providing Level 2
support for IBM storage products. Jon has 29 years of experience
in storage software and management, services, and support, and is
both an IBM Certified IT Specialist and an IBM SAN Certified
Specialist. He is also the UK Chairman of the Storage Networking
Industry Association.
xvi
Preface
xvii
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
xviii
Chapter 1.
Introduction
Computing is based on information. Information is the underlying resource on which all
computing processes are based; it is a company asset. Information is stored on storage
media and is accessed by applications that are running on a server. Often, the information is
a unique company asset. Information is created and acquired every second of every day.
Information is the currency of business.
To ensure that any business delivers the expected results, they must have access to accurate
information, and without delay. The management and protection of business information is
vital for the availability of business processes.
This chapter introduces the concept of a network, storage, and the storage area network
(SAN), which is regarded as the ultimate response to all of these needs.
1.1 Networks
A computer network, often simply called a network, is a collection of computers and devices
that are interconnected by communication channels. These channels allow for the efficient
sharing of resources, services, and information among the network.
Even though this definition is simple, understanding how to make a network work might be
complicated for people who are not familiar with information technology (IT), or who are just
starting out in the IT world. Because of this unfamiliarity, we explain the basic concepts of the
networking world that need to be understood.
For this book, we focus on the Physical, DataLink, Network, and Transport layers.
Chapter 1. Introduction
Local area networks (LANs) are a good place to start. We define LANs as a small or large
network that is limited within the same physical site. This site might be a traditional office or a
corporate building.
In Figure 1-2, you see a basic network where computers and a printer are interconnected by
using physical cables and interconnection devices.
We must keep in mind that any model we choose defines the devices, cables, connectors,
and interface characteristics that we must implement to make it work. We must also support
the protocols for each model layer.
All of the network components are categorized into five groups:
End devices: An end device is a computer system that has a final purpose, such as
desktop computers, printers, storage, or servers.
Network interface: A network interface is an interface between the media and end devices
that can interact with other network interfaces and understands an interconnection model.
Connector: The connector is the physical element at the end of the media that allows a
connection to the network interface.
Media: Media is the physical path that is used to transmit an electrical or optical signal. It
might be wired or wireless, copper, or a fiber optic cable.
Network devices: These devices are used to interconnect multiple end devices as a single
point of interconnection, route communication through separate networks, or provide
network security. Examples of network devices are switches, routers, firewalls, and
directors.
Each network component executes a particular role within a network, and all of them are
required to reach the final goal of making communication possible.
1.3 Storage
To understand what storage is, and because understanding it is a key point for this book, we
start from a basic hard disk drive (HDD). We then progress through to storage systems that
are high performance, fault tolerant, and highly available. During this explanation,
instructional examples are used that might sometimes not reflect reality. However, the basic
examples make it easier to understand for individuals who are just entering the world of
storage systems.
Note: We are aware that solid-state drives (SSD) and Flash arrays form an important part
of any data center today, but for this basic example, we use the HDD as our building block.
Chapter 1. Introduction
RAID 5 uses parity or redundant information. If a block fails, enough parity information is
available to recover the data. The parity information is spread across all of the disks. If a disk
fails, the RAID requires a rebuild and the parity information is used to re-create the lost data.
Figure 1-5 shows this example.
RAID 5 requires a minimum of three disks. In theory, no limitations exist to add disks. This
RAID type combines data safety with the efficient use of disk space. Disk failure does not
result in a service interruption because data is read from parity blocks. RAID 5 is useful for
people who need performance and constant access to their data.
In RAID 5+Spare, disk failure does not require immediate attention because the system
rebuilds itself by using the hot spare. However, the failed disk must be replaced as soon as
possible. A spare disk is an empty disk that is used by the RAID controller only when a disk
fails. Figure 1-6 shows this example.
RAID 5 has better performance for I/O than RAID 1. Depending on the number of disks that
are used to build the RAID, the array disk space utilization is more than two-thirds. RAID 5 is
also managed by a RAID controller that performs the same role as in RAID 1.
Chapter 1. Introduction
Figure 1-7 shows a brief comparison among the most common RAID levels.
RAID types: RAID 1 and RAID 5 are the most common RAID levels. However, many other
levels are not covered in this book. Levels that are not described include RAID 0, 3, 4, and
6; or nested (hybrid) types, such as RAID 0+1 or RAID 5+1. These hybrid types are used in
environments where reliability and performance are key points to be covered from the
storage perspective.
RAID 0
RAID 1
RAID 5
Minimum #
Drive
Features
Data
Protection
No Protection
Single-drive
failure
Single-drive
failure
Read
Performance
High
High
High
Write
Performance
High
Medium
Low
Read
Performance
(degraded)
N/A
Medium
Low
Write
Performance
(degraded)
N/A
High
Low
100%
50%
67% -94%
Capacity
Utilization
Typical
Applications
High-end
workstations,
data logging,
real-time
rendering, very
transitory data
Operating
System,
transaction
databases
Data
warehousing,
web serving,
archiving
Our disk systems seem to be ready to support failures, and they are also high performance.
But what if our RAID controller fails? We might not lose data, but the data is not accessible. Is
a solution available to access this data?
It is almost the same scenario that we initially faced with only one disk as a storage system.
This type of scenario is known as a single point of failure (SPOF). We must add redundancy
by introducing a secondary RAID controller to our storage system.
Now, we are sure that no matter what happens, data is available to be used.
RAID controller role: The RAID controller role in certain cases is performed by the
software. This solution is less expensive than a hardware solution because it does not
require controller hardware; however, it is a slower solution.
We now have several physical HDDs that are managed by two controllers.
Disk pools
When a logical storage volume needs to be provisioned to servers, first the storage RAID
needs to be created. To create the RAID, select the available HDDs and group them together
for a single purpose. The number of grouped HDDs depends on the RAID type that we
choose and the space that is required for provisioning.
To understand what is meant, a basic example is shown that uses the following assumptions:
By issuing these simple steps, we create disk pools. These pools consist of grouping disks
together for a single purpose, such as creating a RAID level, in our case, RAID 5.
In 1.3.2, Redundant Array of Independent Disks on page 6, we mentioned nested (hybrid)
RAIDs, such as 5+0. These solutions are used when the amount of storage data is significant
and important for business continuity. RAID 50 consists of RAID 0 striping across lower-level
RAID 5 arrays. The benefits of RAID 5 are gained while the spanned RAID 0 allows the
incorporation of many more disks into a single logical drive. Up to one drive in each subarray
can fail without data loss.
Chapter 1. Introduction
Also, rebuild times are substantially shorter than the rebuild times of a single large RAID 5
array. See Figure 1-9.
Nested (hybrid) RAIDs: Nested or hybrid RAIDs are a combination of existing RAID levels
that create a RAID to reap the benefits of two separate RAID levels.
This nested RAID 50 can be managed by RC1 or RC2 so we have full redundancy.
Storage systems
We are not far away from building our basic storage system. However, to answer our previous
questions, we need to add two new components and an enclosure.
One of those two components is a CPU that processes all of the required instructions to allow
data to flow. Adding only one CPU creates a single point of failure (SPOF), so we add two
CPUs.
We almost have an independent system. This system must be able to communicate with
other systems in a network. Therefore, a minimum of two network interfaces are required to
be able to avoid a SPOF.
10
Only one step is left. The last step is to put all of these hardware components into an
enclosure. Figure 1-10 shows our storage system.
Chapter 1. Introduction
11
The SAN also eliminates any restriction to the amount of data that a server can access.
Traditionally, a server is limited by the number of storage devices that attach to the individual
server. Instead, a SAN introduces the flexibility of networking to enable one server or many
heterogeneous servers to share a common storage utility. A network might include many
storage devices, including disk, tape, and optical storage. Additionally, the storage utility might
be located far from the servers that it uses.
The SAN can be viewed as an extension to the storage bus concept. This concept enables
storage devices and servers to interconnect by using similar elements, such as LANs and
wide area networks (WANs).
The diagram in Figure 1-11 shows a tiered overview of a SAN that connects multiple servers
to multiple storage systems.
Client
Client
Client
Client
Client
Client
Client
Windows
UNIX
IBM i
Power Systems
System z
DS4000
ESS
Tape
Tape
SANs create new methods of attaching storage to servers. These new methods can enable
great improvements in both availability and performance. The SANs of today are used to
connect shared storage arrays and tape libraries to multiple servers, and they are used by
clustered servers for failover.
A SAN can be used to bypass traditional network bottlenecks. A SAN facilitates direct,
high-speed data transfers between servers and storage devices, potentially in any of the
following three ways:
Server to storage: This method is the traditional model of interaction with storage devices.
The advantage is that the same storage device might be accessed serially or concurrently
by multiple servers.
Server to server: A SAN might be used for high-speed, high-volume communications
between servers.
Storage to storage: This outboard data movement capability enables data to be moved
without server intervention, therefore freeing up server processor cycles for other
activities, such as application processing. Examples include a disk device that backs up its
data to a tape device without server intervention, or a remote device mirroring across the
SAN.
12
SANs allow applications that move data to perform better, for example, by sending data
directly from the source device to the target device with minimal server intervention. SANs
also enable new network architectures where multiple hosts access multiple storage devices
that connect to the same network.
The use of a SAN can potentially offer the following benefits:
Improvements to application availability: Storage is independent of applications and
accessible through multiple data paths for better reliability, availability, and serviceability.
Higher application performance: Storage processing is offloaded from servers and moved
onto a separate network.
Centralized and consolidated storage: Simpler management, scalability, flexibility, and
availability are possible.
Data transfer and vaulting to remote sites: A remote copy of data is enabled for disaster
protection and against malicious attacks.
Simplified centralized management: A single image of storage media simplifies
management.
System z
Windows
UNIX
Power
Systems
IBM i
LINUX
Chapter 1. Introduction
13
IBM participates in many industry standards organizations that work in the field of SANs. IBM
thinks that industry standards must be in place and, if necessary, redefined for SANs to be a
major part of the IT business mainstream.
Probably the most important industry standards organization for SANs is the SNIA. IBM is a
founding member and a board officer in SNIA. The SNIA, other standards organizations, and
IBM are active participants in many of these organizations.
Chapter 1. Introduction
15
16
Chapter 2.
17
18
2.1.2 Requirements
With this scenario in mind, you might consider several requirements for the storage
infrastructures of today:
Unlimited and just-in-time scalability: Businesses require the capability to flexibly adapt to
the rapidly changing demands for storage resources without performance degradation.
System simplification: Businesses require an easy-to-implement infrastructure with a
minimum amount of management and maintenance. The more complex the enterprise
environment, the more costs that are involved in terms of management. Simplifying the
infrastructure can save costs and provide a greater return on investment (ROI).
Flexible and heterogeneous connectivity: The storage resource must be able to support
whatever platforms are within the IT environment. This resource is essentially an
investment protection requirement that allows for the configuration of a storage resource
for one set of systems. It later configures part of the capacity to other systems on an
as-needed basis.
Security: This requirement guarantees that data from one application or system does not
become overlaid or corrupted by other applications or systems. Authorization also requires
the ability to fence off the data of one system from other systems.
Encryption: When sensitive data is stored, it must be read or written only from certain
authorized systems. If for any reason the storage system is stolen, data must never be
available to be read from the system.
Hypervisors: This requirement is for the support of the server, application, and desktop
virtualization hypervisor features for cloud computing.
Speed: Storage networks and devices must be able to manage the high number of
gigabytes and intensive I/O that are required by each business industry.
Availability: This requirement implies both the protection against media failure and the
ease of data migration between devices, without interrupting application processing. This
requirement certainly implies improvements to backup and recovery processes. Attaching
disk and tape devices to the same networked infrastructure allows for fast data movement
between devices, which provides the following enhanced backup and recovery
capabilities:
Serverless backup. This capability is the ability to back up your data without using the
computing processor of your servers.
Synchronous copy. This capability ensures that your data is at two or more places
before your application goes to the next step.
Asynchronous copy. This capability ensures that your data is at two or more places
within a short time. The disk subsystem controls the data flow.
In the following section, we describe the use of SANs as a response to these business
requirements.
19
20
Figure 2-1 illustrates the consolidation movement from the distributed islands of information
toward a single, and, most importantly, simplified infrastructure.
Simplified storage environments have fewer elements to manage. This type of environment
leads to increased resource utilization and simplified storage management. This environment
can provide economies of scale for owning disk storage servers. These environments can be
more resilient and provide an infrastructure for virtualization and automation.
21
2.3.1 Storage
This section briefly describes the major types of storage devices that are available in the
market.
Storage systems
By being contained in a single box, a storage system (hard disk drive (HDD), solid-state drive
(SSD), or Flash) typically has a central control unit that manages all of the I/O. This
configuration simplifies the integration of the system with other devices, such as other disk
systems or servers.
We introduced you to the components of a storage system in Chapter 1, Introduction on
page 1. Depending on the specific functionality that is offered by a particular storage system,
you can make a storage system behave as a small, midsize, or enterprise solution. The
decision about the type of storage system that is more suitable for a SAN implementation
depends on the performance capacity and availability requirements for the particular SAN.
We describe the product portfolio in Chapter 12, IBM Fibre Channel storage area network
product portfolio on page 247.
22
Tape systems
Tape systems, similar to disk systems, are devices that consist of all of the necessary
apparatus to manage the use of tapes for storage. In this case, however, the serial nature of a
tape makes it impossible for them to be treated in parallel. This treatment is because
Redundant Array of Independent Disks (RAID) devices are leading to a simpler architecture
to manage and use.
Three types of tape systems exist: drives, autoloaders, and libraries. An overview of each
type of system is provided.
Tape drives
As with disk drives, tape drives are the means by which tapes can connect to other devices.
They provide the physical and logical structure for reading from, and writing to tapes.
Tape autoloaders
Tape autoloaders are autonomous tape drives that can manage tapes and perform automatic
backup operations. They are typically connected to high-throughput devices that require
constant data backup.
Tape libraries
Tape libraries are devices that can manage multiple tapes simultaneously. They can be
viewed as a set of independent tape drives or autoloaders. They are typically deployed in
systems that require massive storage capacity, or that need a type of data separation that
results in multiple single-tape systems. Because a tape is not a random-access media, tape
libraries cannot provide parallel access to multiple tapes as a way to improve performance.
However, they can provide redundancy as a way to improve data availability and
fault-tolerance.
The circumstances under which each of these systems, or even a disk system, might be used
strongly depends on the specific requirements of a particular SAN implementation. However,
disk systems are used for online storage because of their superior performance. Tape
systems are ideal for offline, high-throughput storage, because of the lower cost of storage
per byte.
The next section describes the prevalent connectivity interfaces, protocols, and services for
building a SAN.
23
The SCSI bus is a parallel bus, which comes in several variants (Figure 2-2).
Fibre Channel: For more information about parallel and serial data transfer, see
Chapter 3, Fibre Channel internals on page 33.
24
Figure 2-3 shows the evolution of Fibre Channel speeds. Fibre Channel is described in
greater depth throughout this publication.
Figure 2-4 shows an overview of the Fibre Channel model. The diagram shows the Fibre
Channel, which is divided into four lower layers (FC-0, FC-1, FC-2, and FC-3) and one upper
layer (FC-4). FC-4 is where the upper-level protocols are used, such as SCSI-3, Internet
Protocol (IP), and Fibre Channel connection (FICON).
25
Lower-level layers
As Figure 2-5 shows, only three stacks can directly interact with the physical wire: Ethernet,
SCSI, and Fibre Channel. Because of this configuration, these models are considered the
lower-level layers. All of the other stacks are combinations of the layers, such as Internet
SCSI (iSCSI), Fibre Channel over IP (FCIP), and Fibre Channel over Ethernet (FCoE), which
are also called the middle-level layers.
We assume that you have a basic knowledge of Ethernet, which is typically used on
conventional server-to-server or workstation-to-server network connections. The connections
build up a common-bus topology by which every attached device can communicate with every
other attached device by using this common bus. Ethernet speed is increasing as it becomes
more pervasive in the data center. Key concepts of Ethernet are describer later in this book.
Middle-level layers
This section consists of the transport protocol and session layers.
Fibre Channel over Ethernet (FCoE): FCoE is described later in this book. It is a vital
model for the Converged Network Adapter (CNA).
iSCSI enables the implementation of IP-based SANs, enabling clients to use the same
networking technologies, for both storage and data networks. Because iSCSI uses TCP/IP,
iSCSI is also suited to run over almost any physical network. By eliminating the need for a
second network technology just for storage, iSCSI has the potential to lower the costs of
deploying networked storage.
Higher-level layers
This section consists of the presentation and application layers.
Server-attached storage
The earliest approach was to tightly couple the storage device with the server. This
server-attached storage approach keeps performance overhead to a minimum. Storage is
attached directly to the server bus by using an adapter, and the storage device is dedicated to
a single server. The server itself controls the I/O to the device, issues the low-level device
commands, and monitors device responses.
Chapter 2. Storage area networks
27
Initially, disk and tape storage devices had no onboard intelligence. They merely ran the I/O
requests of the server. The subsequent evolution led to the introduction of control units (CUs).
These units are storage offload servers that contain a limited level of intelligence. The CUs
can perform functions, such as I/O request caching for performance improvements or dual
copying data (RAID 1) for availability. Many advanced storage functions are developed and
implemented inside the CU.
Network-attached storage
Network-attached storage (NAS) is basically a LAN-attached file server that serves files by
using a network protocol, such as Network File System (NFS). NAS refers to storage
elements that connect to a network and provide file access services to computer systems. An
NAS storage element consists of an engine that implements the file services (by using access
protocols, such as NFS or Common Internet File System (CIFS)) and one or more devices, on
which data is stored. NAS elements might be attached to any type of network.
From a SAN perspective, a SAN-attached NAS engine is treated just like any other server.
However, NAS does not provide any of the activities that a server in a server-centric system
typically provides, such as email, authentication, or file management.
NAS allows more hard disk storage space to be added to a network that already uses
servers, without shutting them down for maintenance and upgrades. With an NAS device,
storage is not a part of the server. Instead, in this storage-centric design, the server still
handles all of the processing of the data, but an NAS device delivers the data to the user.
An NAS device does not need to be located within the server, but an NAS device can exist
anywhere in the LAN. An NAS device can consist of multiple networked NAS devices. These
units communicate to a host by using Ethernet and file-based protocols. This method is in
contrast to the disk units that are already described, which use Fibre Channel Protocol (FCP)
and block-based protocols to communicate.
NAS storage provides acceptable performance and security, and it is often less expensive for
servers to implement (for example, Ethernet adapters are less expensive than Fibre Channel
adapters).
To bridge the two worlds and open up new configuration options for clients, certain vendors,
including IBM, sell NAS units that act as a gateway between IP-based users and
SAN-attached storage. This configuration allows the connection of the storage device and
shares the storage device between your high-performance database servers (attached
directly through FC) and your users (attached through IP). These users do not have strict
performance requirements.
NAS is an ideal solution for serving files that are stored on the SAN to users in cases where it
is impractical and expensive to equip users with Fibre Channel adapters. NAS allows those
users to access your storage through the IP-based network that they already have.
2.3.3 Servers
Each server platform (IBM z Systems, UNIX, IBM AIX, HP-UX, Sun Solaris, Linux, IBM i,
and Microsoft Windows Server) implements SAN solutions by using various interconnections
and storage technologies. The following sections review these solutions and the
implementation on each platform.
28
Mainframe servers
A mainframe is a single, monolithic, and possibly multiple processor, high-performance
computer system. Apart from the fact that the IT evolution is pointing toward a more
distributed and loosely coupled infrastructure, mainframes still play an important role in
businesses that depend on massive storage capabilities.
The IBM System z is a processor and operating system mainframe set. Historically, the IBM
z Systems servers supported many operating systems, such as z/OS, IBM OS/390, VM,
VSE, and Transaction Processing Facility (TPF), which were enhanced over the years. The
processor to storage device interconnection also evolved from a bus and tag interface to
ESCON channels, and now to FICON channels. Figure 2-6 shows the various
processor-to-storage interfaces.
Because of architectural differences, and strict data integrity and management requirements,
the implementation of FICON is somewhat behind that of FCP on open systems. However, at
the time of writing this book, FICON caught up with FCP SANs, and they coexist amicably.
For the latest news about IBM z Systems FICON connectivity, see this website:
http://www.ibm.com/systems/z/hardware/connectivity/index.html
In addition to FICON for traditional z Systems operating systems, IBM has standard Fibre
Channel adapters for use with z Systems servers that can implement Linux.
29
UNIX servers
Originally designed for high-performance computer systems, such as mainframes, todays
UNIX operating systems appear on many hardware platforms, ranging from Linux personal
computers to dedicated large-scale workstations. Because of the popularity and maturity of
UNIX, UNIX also plays an important role on both existing and earlier IT infrastructures.
The IBM Power Systems servers run a UNIX operating system that is called AIX. The
Power Systems servers offer various processor-to-storage interfaces, including SCSI,
serial-attached SCSI (SAS), and Fibre Channel. The Serial Storage Architecture (SSA)
interconnection is primarily used for disk storage. Fibre Channel adapters can connect to tape
and disk. Figure 2-7 shows the various processor-to-storage interconnection options for the
Power Systems family.
The various UNIX system vendors in the market deploy variations of the UNIX operating
system. Each product offers unique enhancements. The various vendors UNIX operation
system implementations often support separate file systems, such as the journaled file
system (JFS), enhanced journaled file system (JFS2), and the IBM Andrew File System
(AFS). The server-to-storage interconnection is similar to Power Systems, as shown in
Figure 2-7.
For the latest IBM Power Systems storage products, see this website:
http://www.ibm.com/systems/storage/product/power.html
30
Single-level storage
Single-level storage (SLS) is probably the most significant differentiator in a SAN solution
implementation on an IBM i server. This IBM i differentiator is a factor when compared to
other systems such as z/OS, UNIX, and Windows. In IBM i, both the major storage (memory)
and the auxiliary storage (disks) are treated as a large virtual address space that is known as
SLS.
Figure 2-8 compares the IBM i SLS addressing with the way that Windows or UNIX systems
work, by using the processor local storage. With 32-bit addressing, each process (job) has
4 GB of addressable memory. With 64-bit SLS addressing, over 18 million terabytes
(18 exabytes) of addressable storage are possible. Because a single page table maps all
virtual addresses to physical addresses, task switching is efficient. SLS further eliminates the
need for address translation, therefore speeding up data access.
Figure 2-8 IBM i versus 32-bit Windows Server or UNIX storage addressing
The IBM i SAN support was rapidly expanded. IBM i servers now support attachment to
switched fabrics and to most of the IBM SAN-attached storage products.
For more information, see this IBM i SAN website:
http://www.ibm.com/systems/i/hardware/storage/
31
32
Chapter 3.
33
34
Scalability limitations
The amount of data that is available to the server is determined by the number of devices that
can attach to the bus. The amount is also determined by the number of buses that are
attached to the server. Up to 15 devices can be attached to a server on a single SCSI bus. In
practice, because of performance limitations due to arbitration, commonly no more than four
or five devices are attached in this way. This factor limits the scalability in terms of the number
of devices that can connect to the server.
35
A SCSI propagation delay in sending data in parallel along multiple lines leads to a
phenomenon that is known as skew. Skew means that all bits might not arrive at the target
device at the same time. Figure 3-2 shows this result.
Driver Skew
Cable Skew
Receiver Skew
Data 1
Driver
Receiver
Data 2
Driver
Receiver
Data 3
Driver
Receiver
Data 4
Driver
Receiver
Clock
Driver
Receiver
Arrival occurs during a small window of time, depending on the transmission speed and the
physical length of the SCSI bus. The need to minimize the skew limits the distance that
devices can be positioned away from the initiating server to 2 meters (6.5 ft) - 25 meters
(82 ft). The distance depends on the cycle time. Faster speed means shorter distance.
36
Distance limitations
The distances refer to the maximum length of the SCSI bus, including all attached devices.
Figure 3-3 shows the SCSI distance limitations. These limitations might severely restrict the
total GB capacity of the disk storage that can be attached to an individual server.
20 MB/sec
Device
Termination
.....
Up to 25 meters
40 MB/sec
Device
Termination
.....
Up to 12 meters
80 MB/sec
Device
Termination
.....
Up to 12 meters
Ultra2 SCSI
Device sharing
Many applications require the system to access several devices, or for several systems to
share a single device. SCSI can enable this sharing by attaching multiple servers or devices
to the same bus. This structure is known as a multi-drop configuration.
37
SCSI Initiator
Driver
Receiver
Terminator
Data Bus
Control Signals
Driver
Receiver
Driver
Receiver
Driver
Receiver
Driver
Receiver
Disk Drive
Disk Drive
Disk Drive
Disk Drive
To avoid signal interference, and therefore possible data corruption, all unused ports on a
parallel SCSI bus must be terminated correctly. Incorrect termination can result in transaction
errors or failures.
Normally, only a single server can access data on a specific disk with a SCSI bus. In a shared
bus environment, it is clear that all devices cannot transfer data at the same time. SCSI uses
an arbitration protocol to determine the device that can gain access to the bus. Arbitration
occurs before and after every data transfer on the bus. While arbitration takes place, no data
movement can occur. This loss of movement represents an additional performance overhead
that reduces bandwidth utilization, substantially reducing the effective data rate that is
achievable on the bus. Actual rates are typically less than 50% of the rated speed of the SCSI
bus.
It is clear that the physical parallel SCSI bus architecture has several significant speed,
distance, and availability limitations. These limits make it increasingly less suitable for many
applications in todays networked IT infrastructure. However, the SCSI protocol is deeply
embedded in the way that commonly encountered operating systems handle user requests
for data. Therefore, requiring a move to new protocols is a major inhibitor to progress.
38
Fibre Channel is a multiple layer network that is based on a series of American National
Standards Institute (ANSI) standards that define characteristics and functions for moving data
across the network. These standards include the definitions of physical interfaces, for
example:
Like other networks, information is sent in structured packets or frames, and data is serialized
before transmission. But, unlike other networks, the Fibre Channel architecture includes
significant hardware processing to deliver high performance.
Fibre Channel uses a serial data transport scheme that is similar to other computer networks,
which stream packets (frames) of bits, one behind the other, in a single data line to achieve
high data rates.
Serial Transfer
Target
Initiator
Parallel Transfer
Target
Initiator
Serial transfer does not suffer from the problem of skew, so speed and distance are not
restricted in the same way that parallel data transfers are restricted. Figure 3-5 shows the
process of parallel data transfers versus serial data transfers.
Serial transfer enables simpler cabling and connectors, and also the routing of information
through switched networks. Fibre Channel can operate over longer distances, both natively
and by implementing cascading, and longer with the introduction of repeaters. Just as LANs
can be interlinked in wide area networks (WANs) by using high-speed gateways, campus
SANs can be interlinked to build enterprise-wide SANs.
39
Whatever the topology, information is sent between two nodes, which are the source
(transmitter or initiator) and destination (receiver or target). A node is a device, such as a
server (personal computer, workstation, or mainframe) or peripheral device, such as a disk or
tape drive, or a video camera. Frames of information are passed between nodes, and the
structure of the frame is defined by a protocol. Logically, a source and target node must use
the same protocol, but each node might support several protocols or data types.
Therefore, Fibre Channel architecture is flexible in its potential application. Fibre Channel
transport layers are protocol independent, enabling the transmission of multiple protocols.
Using a credit-based flow control approach, Fibre Channel can deliver data as fast as the
destination device buffer can receive it. And low transmission overhead enables high
sustained utilization rates without the loss of data.
Therefore, Fibre Channel combines the best characteristics of traditional I/O channels with
the characteristics of computer networks:
High performance for large data transfers by using simple transport protocols and
extensive hardware assists
Serial data transmission
A physical interface with a low error rate definition
Reliable transmission of data with the ability to guarantee or confirm error-free delivery of
the data
The ability to package data in packets (frames, in Fibre Channel terminology)
Flexibility in terms of the types of information that can be transported in frames (such as
data, video, and audio)
Use of existing device-oriented command sets, such as SCSI and FCP
A vast expansion in the number of devices that can be addressed when compared to I/O
interfaces: a theoretical maximum of more than 15 million ports
Several factors make the Fibre Channel architecture ideal for the development of enterprise
SANs. One example is the high degree of flexibility, availability, and scalability of the
architecture. Other factors include the combination of multiple protocols at high speeds over
long distances, and the broad acceptance of the Fibre Channel standards by vendors
throughout the IT industry.
The following topics describe several key concepts that are mentioned in the previous pages
and that are behind Fibre Channel SAN implementations. We also introduce more Fibre
Channel SAN terminology and jargon that you can expect to encounter.
3.2 Layers
Fibre Channel (FC) is broken up into a series of five layers. The concept of layers, starting
with the International Organization for Standardization/open systems interconnection
(ISO/OSI) seven-layer model, allows the development of one layer to remain independent of
the adjacent layers. Although a Fibre Channel contains five layers, those layers follow the
general principles that are stated in the ISO/OSI model.
The series of five layers that make up a Fibre Channel can be categorized into the following
layers:
Physical and signaling layer
Upper layer
40
Fibre Channel is a layered protocol. Figure 3-6 shows the upper and physical layers.
Application
FC-4
FC-3
Common Services
FC-2
Signaling Protocol
(Framing/Flow Control)
FC-1
Transmission Protocol
(Encode/Decode)
FC-0
Physical Interface/Media
Upper
Layers
Physical
Layers
Laser safety
Fibre Channel often uses lasers to transmit data, and can, therefore, present an optical health
hazard. The FC-0 layer defines an open fiber control (OFC) system, and it acts as a safety
interlock for point-to-point fiber connections that use semiconductor laser diodes as the
optical source. If the fiber connection is broken, the ports send a series of pulses until the
physical connection is re-established and the necessary handshake procedures are followed.
41
Data is transmitted in 4-byte ordered sets that contain data and control characters. Ordered
sets provide the availability to obtain bit and word synchronization, which also establishes
word boundary alignment.
Together, FC-0, FC-1, and FC-2 form the FC-PH.
Upper layers
The upper layer includes two layers: FC-3 and FC-4.
Multicast delivers a single transmission to multiple destination ports. This method includes
the ability to broadcast to all nodes or a subset of nodes.
Upper-layer protocol mapping: FC-4
The highest layer, FC-4, provides the application-specific protocols. Fibre Channel is equally
adept at transporting both the network and channel information and allows both protocol
types to be transported concurrently over the same physical interface.
Through mapping rules, a specific FC-4 describes how upper-layer protocol (ULP) processes
of the same FC-4 type interoperate.
A channel example is FCP. This protocol is used to transfer SCSI data over Fibre Channel. A
networking example is sending IP packets between the nodes. FICON is another ULP in use
today for mainframe systems. FICON is a contraction of Fibre Connection and refers to
running ESCON traffic over Fibre Channel.
42
3.3.1 Attenuation
The pulse is weaker because all glass absorbs light. More accurately, impurities in the glass
can absorb light but the glass itself does not absorb light at the wavelengths of interest. In
addition, variations in the uniformity of the glass cause the scattering of the light. Both the rate
of light absorption and the amount of scattering depend on the wavelength of the light and the
characteristics of the particular glass. Most light loss in a modern fiber is caused by
scattering.
Polarization
Conventional communication optical fiber is cylindrically symmetric, but it contains
imperfections. Light traveling down such a fiber is changed in polarization. (In current optical
systems, this change does not matter, but in future systems, it might become a critical issue.)
43
Dispersion
Dispersion occurs when a pulse of light is spread out during transmission on the fiber. A short
pulse becomes longer and ultimately joins with the pulse behind, making the recovery of a
reliable bit stream impossible. (In most communications systems, bits of information are sent
as pulses of light: 1 = light, 0 = dark. But even in analog transmission systems where
information is sent as a continuous series of changes in the signal, dispersion causes
distortion.) Many kinds of dispersion exist, and each kind works differently. The three most
important kinds of dispersion are described.
Modal dispersion
When you use a multimode fiber, the light can take many paths or modes as it travels within
the fiber. The distance that is traveled by light in each mode differs from the distance that is
traveled in other modes. When a pulse is sent, parts of that pulse (rays or quanta) take many
different modes (usually all available modes). Therefore, certain components of the pulse
arrive before other components of the pulse. The difference between the arrival time of light
that takes the fastest mode, versus the arrival time of light that takes the slowest mode,
obviously becomes greater as the distance becomes greater.
Waveguide dispersion
Waveguide dispersion is a complex effect, and it is caused by the shape and index profile of
the fiber core. However, this effect can be controlled by careful design and, in fact, waveguide
dispersion can be used to counteract material dispersion.
Noise
One of the great benefits of fiber optical communications is that the fiber does not pick up
noise from outside the system. However, various kinds of noise can come from components
within the system. Mode partition noise can be a problem in single-mode fiber, and modal
noise is a phenomenon in multimode fiber.
It is not our intention to delve any deeper into optical than the information that is already
described.
44
Single-Mode Fiber
Cladding (125 um)
Core ( 9 um)
Multi-Mode Fiber
Cladding (125 um)
Core (50 um or 62.5 um)
45
46
Table 3-1 lists the cable types and their speed and distance.
Table 3-1 Fibre Channel modes, speeds, and distances
Fiber mode
Speed (MBps)
Transmitter
Medium
Distance
Single-mode fiber
1600
1310 nm
longwave light
1600-SM-LC-L
0.5 m - 10 km
1490 nm
longwave light
1600-SM-LZ-I
0.5 m - 2 km
1310 nm
longwave light
800-SM-LC-L
2 m - 10 km
800-SM-LC-I
2 m - 1.4 km
1310 nm
longwave light
400-SM-LC-L
2 m - 10 km
400-SM-LC-M
2 m - 4 km
400-SM-LL-I
2 m - 2 km
1550 nm
longwave light
200-SM-LL-V
2 m - 50 km
1310 nm
longwave light
200-SM-LC-L
2 m - 10 km
200-SM-LL-I
2 m - 2 km
1550 nm
longwave light
100-SM-LL-V
2 m - 50 km
1310 nm
longwave light
100-SM-LL-L
2 m - 10 km
100-SM-LC-L
2 m - 10 km
100-SM-LL-I
2 m - 2 km
800
400
200
100
47
Fiber mode
Speed (MBps)
Transmitter
Medium
Distance
Multi-mode fibera
1600
850 nm
shortwave light
1600-M5F-SN-I
0.5 m - 125 m
1600-M5E-SN-I
0.5 - 100 m
1600-M5-SN-S
0.5 - 35 m
1600-M6-SN-S
0.5 - 15 m
800-M5F-SN-I
0.5 - 190 m
800-M5E-SN-I
0.5 - 150 m
800-M5-SN-S
0.5 - 50 m
800-M6-SN-S
0.5 - 21 m
400-M5F-SN-I
0.5 - 400 m
400-M5E-SN-I
0.5- 380 m
400-M5-SN-I
0.5 - 150 m
400-M6-SN-I
0.5- 70 m
200-M5E-SN-I
0.5 - 500 m
200-M5-SN-I
0.5 - 300 m
200-M6-SN-I
0.5 - 150 m
100-M5E-SN-I
0.5 - 860 m
100-M5-SN-I
0.5 - 500 m
100-M6-SN-I
0.5 - 300 m
100-M5-SL-I
2 - 500 m
100-M6-SL-I
2 - 175 m
800
400
200
100
Table 3-2 shows the MMF designations, optical multi-mode (OM) numbering, fiber-optic cable
diameters, and FC media designation.
Table 3-2 Optical multimode designations
48
Multi-mode fiber
FC media designation
OM1
62.5 m
M6
OM2
50 m
M5
OM3
50 m
M5E
OM4
50 m
M5F
Description
Requires an
acknowledgment
Yes
Yes
No
Yes
Yes
Switch-to-switch communication
Yes
49
3.4.1 Class 1
In class 1 service, a dedicated connection source and destination are established through the
fabric during the transmission. Class 1 service provides acknowledged service. This class of
service ensures that the frames are received by the destination device in the same order in
which they are sent. This class reserves full bandwidth for the connection between the two
devices. It does not provide for a good utilization of the available bandwidth because it blocks
another possible contender for the same device. Because of this blocking and the necessary
dedicated connections, class 1 is rarely used.
3.4.2 Class 2
Class 2 is a connectionless, acknowledged service. Class 2 makes better use of available
bandwidth because it allows the fabric to multiplex several messages on a frame-by-frame
basis. While frames travel through the fabric, they can take separate routes, so class 2
service does not guarantee in-order delivery. Class 2 relies on upper-layer protocols to take
care of the frame sequence. The use of acknowledgments reduces available bandwidth,
which needs to be considered in large-scale busy networks.
3.4.3 Class 3
No dedicated connection is available in class 3, and the received frames are not
acknowledged. Class 3 is also called datagram connectionless service. It optimizes the use of
fabric resources, but it is now up to the upper-layer protocol to ensure that all frames are
received in the correct order. The upper-layer protocol also needs to request to the source
device the retransmission of missing frames. Class 3 is a commonly used class of service in
Fibre Channel networks.
3.4.4 Class 4
Class 4 is a connection-oriented service, which is similar to class 1. The major difference is
that class 4 allocates only a fraction of the available bandwidth of the path through the fabric
that connects two N_ports. Virtual circuits (VCs) are established between two N_ports with
guaranteed quality of service (QoS), including bandwidth and latency. Like class 1, class 4
guarantees the in-order delivery of frames and provides an acknowledgment of delivered
frames. However, now the fabric is responsible for multiplexing frames of different VCs. Class
4 service is intended for multimedia applications, such as video, and for applications that
allocate an established bandwidth by department within the enterprise. Class 4 is included in
the FC-PH-2 standard.
3.4.5 Class 5
Class 5 is called isochronous service, and is intended for applications that require immediate
delivery of the data as it arrives, with no buffering. Class 5 is not clearly defined yet, and it is
not included in the FC-PH documents.
50
3.4.6 Class 6
Class 6 is a variant of class 1, and it is known as a multicast class of service. It provides
dedicated connections for a reliable multicast. An N_port might request a class 6 connection
for one or more destinations. A multicast server in the fabric establishes the connections,
receives the acknowledgment from the destination ports, and sends the acknowledgment
back to the originator. When a connection is established, the connection is retained and
guaranteed by the fabric until the initiator ends the connection. Class 6 was designed for
applications, such as audio and video, that require multicast functionality. Class 6 is included
in the FC-PH-3 standard.
3.4.7 Class F
Class F service is defined in the Fibre Channel Switched Fabric (FC-SW) standard and the
FC-SW-2 standard for use by switches that communicate through inter-switch links (ISLs). It
is a connectionless service with notification of non-delivery between E_ports that are used for
the control, coordination, and configuration of the fabric. Class F is similar to class 2. The
major difference is that class 2 works with N_ports that send data frames. Class F is used by
E_ports for the control and management of the fabric.
51
25
Notation D
5b/6b Encoder
3b/4b Encoder
10-bit characters
The following steps occur in the encoding example that is shown in Figure 3-8:
1.
2.
3.
4.
52
Running disparity
The conversion of the 8-bit data bytes results in two 10-bit results. The encoder must choose
one of these results to use. This decision is achieved by monitoring the running disparity of
the previously processed character. For example, if the previous character showed a positive
disparity, the next character that is issued might have an encoded value that represents
negative disparity.
In the example that is used in Figure 3-8 on page 52, the encoded value, when the running
disparity is either positive or negative, is the same. This outcome is legitimate. In certain
cases, the encoded value differs, and in other cases, the encoded value is the same.
In Figure 3-8 on page 52, the encoded 10-bit byte has 5 bits that are set and 5 bits that are
unset. The only possible results of the 8b/10b encoding are shown:
If 5 bits are set, the byte is said to have neutral disparity.
If 4 bits are set and 6 bits are unset, the byte is said to have negative disparity.
If 6 bits are set and 4 bits are unset, the byte is said to have positive disparity.
The rules of Fibre Channel define that a byte that is sent cannot take the positive or negative
disparity above one unit. Therefore, if the current running disparity is negative, the next byte
that is sent must have one of these properties:
Neutral disparity:
Keeping the current running disparity negative.
The subsequent byte needs to have either neutral or positive disparity.
Positive disparity:
Making the new current running disparity neutral.
The subsequent byte has either positive, negative, or neutral disparity.
Number of bits: At any time or at the end of any byte, the number of set bits and unset bits
that passes over a Fibre Channel link differ only by a maximum of two.
K28.5
In addition to the fact that many 8-bit numbers encode to two 10-bit numbers under the
8b/10b encoding scheme, other key features exist.
Certain 10-bit numbers cannot be generated from any 8-bit number. Therefore, it is not
possible to see these particular 10-bit numbers as part of a flow of data. This outcome is
useful because these particular 10-bit numbers can be used by the protocol for signaling or
control.
These characters are referred to as comma characters. Instead of the prefix D, they use the
prefix K.
The only one character that gets used in Fibre Channel is the one that is known as K28.5,
and it has a special property.
Table 3-4 shows the two 10-bit encodings of K28.5.
Table 3-4 10-bit encoding of K28.5
Name of character
K28.5
Positive
001111 1010
110000 0101
53
All of the 10-bit bytes that are possible by using the 8b/10b encoding scheme have either 4, 5,
or 6 bits that are set. The K28.5 character is special because it is the only character that is
used in Fibre Channel that has 5 consecutive bits that are set or unset. All other characters
have four or fewer consecutive bits of the same setting.
When you determine the significance of the bit settings, consider two things:
The 1s and 0s actually represent light and dark on the fiber (assuming fiber optic medium).
A 010 pattern effectively is a light pulse between two periods of darkness. A 0110
represents the same, except that the pulse of light lasts for twice the length of time.
Because the two devices have their own clocking circuitry, the number of consecutive set
bits, or consecutive unset bits, becomes important. For example, device 1 is sending to
device 2, and the clock on device 2 is running 10% faster than the clock on device 1. If
device 1 sent 20 clock cycles worth of set bits, device 2 counts 22 set bits. (This example
is provided merely to illustrate the point.) The worst possible case that you can have in
Fibre Channel is 5 consecutive bits of the same setting within 1 byte: the K28.5.
Because K28.5 is the only character with five consecutive bits of the same setting, Fibre
Channel hardware can look out for it specifically. Because K28.5 is used for control, this
setting is useful. This setting allows the hardware to be designed for maximum efficiency.
64b/66b encoding
Communications of 10 Gbps and 16 Gbps use 64/66b encoding. Sixty-four bits of data are
transmitted as a 66-bit entity. The 66-bit entity is made by prefixing one of two possible 2-bit
preambles to the 64 bits to be transmitted. If the preamble is 01, the 64 bits are entirely data.
If the preamble is 10, an 8-bit type field follows, plus 56 bits of control information and data.
The preambles 00 and 11 are not used, and they generate an error, if seen.
The use of the 01 and 10 preambles guarantees a bit transmission every 66 bits, which
means that a continuous stream of 0s or 1s cannot be valid data. It also allows easier clock
and timer synchronization because a transmission must be seen every 66 bits.
The overhead of the 64B/66B encoding is considerably lower than the more common 8b/10b
encoding scheme.
54
The frame delimiters, the start-of-frame (SOF), and end-of-frame (EOF) ordered sets
establish the boundaries of a frame. They immediately precede or follow the contents of a
frame. Eleven types of SOF and eight types of EOF delimiters are defined for the fabric and
N_port sequence control.
The two primitive signals, idle and receiver ready (R_RDY), are ordered sets that are
designated by the standard to have a special meaning. An idle is a primitive signal that is
transmitted on the link to indicate that an operational port facility is ready for frame
transmission and reception. The R_RDY primitive signal indicates that the interface buffer is
available for receiving further frames.
A primitive sequence is an ordered set that is transmitted and repeated continuously to
indicate specific conditions within a port. Or, the set might indicate conditions that are
encountered by the receiver logic of a port. When a primitive sequence is received and
recognized, a corresponding primitive sequence or idle is transmitted in response.
Recognition of a primitive sequence requires the consecutive detection of three instances of
the same ordered set. The primitive sequences that are supported by the standard include
the following settings:
Offline state (OLS)
The offline primitive sequence is transmitted by a port to indicate one of the following
conditions:
The port is beginning the link initialization protocol.
The port received and recognized the NOS protocol.
The port is entering the offline status.
Not operational (NOS)
The not operational primitive sequence is transmitted by a port in a point-to-point or fabric
environment to indicate that the transmitting port detected a link failure. Or, the NOS might
indicate an offline condition that is waiting for the OLS sequence to be received.
Link reset (LR)
The link reset primitive sequence is used to initiate a link reset.
Link reset response (LRR)
Link reset response is transmitted by a port to indicate that it recognizes a link reset
sequence and performed the correct link reset.
Data transfer
To send data over Fibre Channel, though, we need more than merely the control
mechanisms. Data is sent in frames. One or more related frames make up a sequence. One
or more related sequences make up an exchange.
3.6.2 Frames
Fibre Channel places a restriction on the length of the data field of a frame at 528
transmission words, which is 2112 bytes. See Table 3-5 on page 56. Larger amounts of data
must be transmitted in several frames. This larger unit that consists of multiple frames is
called a sequence. An entire transaction between two ports is made up of sequences that are
administered by an even larger unit that is called an exchange.
55
Frame arrival: Certain classes of Fibre Channel communication guarantee that the frames
arrive at the destination in the same order in which they were transmitted. Other classes do
not. If the frames arrive in the same order in which they were sent, the delivery is
considered an in order delivery of frames.
A frame consists of the following elements:
SOF delimiter
Frame header
Optional headers and payload (data field)
Cyclic redundancy check (CRC) field
EOF delimiter
32 bit
24 bit
24 bit
start of dest port src port
frame address address
Word 0 Word 1 Word 2
SOF
Control
information
Words 3-6
24 byte FC header
Frame payload
(0 - 2112 bytes)
SCSI data
CRC
CRC
End of
frame
delimiter
EOF
Framing rules
The following rules apply to the framing protocol:
A frame is the smallest unit of information transfer.
A sequence has at least one frame.
An exchange has at least one sequence.
Transmission word
A transmission word is the smallest transmission unit that is defined in Fibre Channel. This
unit consists of four transmission characters, 4 x 10 or 40 bits. When information that is
transferred is not an even multiple of 4 bytes, the framing protocol adds fill bytes. The fill bytes
are stripped at the destination.
Frames are the building blocks of Fibre Channel. A frame is a string of transmission words
that are prefixed by a SOF delimiter and followed by an EOF delimiter. Table 3-5 shows the
way that transmission words make up a frame.
Table 3-5 Transmission words in a frame
SOF
Frame header
CRC
EOF
1 TW
6 TW
0-528 TW
1 TW
1 TW
Frame header
Each frame includes a header that identifies the source and destination of the frame. The
frame also includes control information that manages the frame, sequences, and exchanges
that are associated with that frame.
56
Byte 1
Byte 2
Word 0
R_CTL
Destination_ID (D_ID)
Word 1
Reserved
Source_ID (S_ID)
Word 2
Type
Word 3
SEQ_ID
DF_CTL
Word 4
Word 5
Parameter
Byte 3
3.6.3 Sequences
The information in a sequence moves in one direction from a source N_port to a destination
N_port. Various fields in the frame header are used to identify the beginning, middle, and end
of a sequence. Other fields in the frame header are used to identify the order of frames in
case they arrive out of order at the destination.
3.6.4 Exchanges
Two other fields of the frame header identify the exchange ID. An exchange is responsible for
managing a single operation that might span several sequences, possibly in opposite
directions. The source and destination can have multiple exchanges active at a time.
57
Using SCSI as an example, a SCSI task is an exchange. The SCSI task is made up of one or
more information units. The following information units (IUs) are relevant for this SCSI task:
Command IU
Transfer ready IU
Data IU
Response IU
Each IU is one sequence of the exchange. Only one participant sends a sequence at a time.
Figure 3-10 shows the flow of the exchange, sequence, and frames.
Figure 3-10 Fibre Channel (FC) exchange, sequence, and frame flow
58
3.6.6 Latency
The term latency means the delay between an action that is requested and the action
occurring.
Latency occurs almost everywhere because it takes time and energy to perform an action.
The following areas highlight where you need to be aware of the latency in a SAN:
Ports
Switches and directors
Inter-Chassis Links in a DCX director
Long-distance links
Inter-Switch Links
Application-specific integrated circuits (ASICs)
59
3.7.4 Performance
You can see the importance of the number of buffers in overall performance. You need
enough buffers to ensure that the transmitting port can continue sending frames without
stopping to be able to use the full bandwidth. Using sufficient buffers is important with
distance. At 1 Gbps, a frame occupies about 75 m (246 ft) - 4 km (2.48 miles) of fiber, which
depends on the size of the data payload. In a 100 km (62 miles) link, many frames can be
sent before the first frame reaches its destination. You need an ACK back to start replenishing
the EE_Credit or an R_RDY indication to replenish the BB_Credit.
60
For a moment, consider frames with 2 KB of data. These frames occupy approximately 4 km
(2.48 miles) of fiber. You are able to send about 25 frames before the first frame arrives at the
far end of the 100 km (62 miles) link. You are able to send another 25 frames before the first
R_RDY or ACK indication is received. Therefore, you need at least 50 buffers to allow for
non-stop transmission at a 100 km (62 miles) distance with frames of this size. If the frame
size is reduced, more buffers are required to allow non-stop transmission. In brief, the buffer
credit management is critical in long-distance communication. Therefore, the correct buffer
credit allocation is important to obtain optimal performance. Incorrect allocation of the buffer
credit might result in a delay of transmission over the Fibre Channel link. As a preferred
practice, always refer to the default buffer and maximum buffer credit values for each model of
switch from each vendor.
61
62
Chapter 4.
63
4.1 Ethernet
In Chapter 1, Introduction on page 1, we briefly introduced the network and the importance
of the models. The Ethernet standard fits into layer 2 of the open systems interconnection
(OSI) model. The standard refers to the media access layer that devices are connected to
(the cable) and compete for access by using the Carrier Sense Multiple Access with Collision
Detection (CSMA/CD) protocol.
64
The Ethernet frame that is shown in Figure 4-1 contains the following components:
Preamble: A preamble is a stream of bits that are used to allow the transmitter and
receiver to synchronize their communication. The preamble is an alternating pattern of 56
binary ones and zeros. The preamble is immediately followed by the Start Frame
Delimiter.
Start Frame Delimiter: The Start Frame Delimiter is always 10101011. It is used to
indicate the beginning of the frame information.
Destination Media Access Control: The destination Media Access Control (MAC) is the
address of the system that is receiving data. When a NIC is listening to the wire, the NIC is
checking this field for its own MAC address.
Source Media Access Control: The source Media Access Control (MAC) is the MAC
address of the system that is transmitting data.
Length: This field is the length of the entire Ethernet frame, in bytes. Although this field can
hold any value 0 - 65,534, it is rarely larger than 1500. This smaller value is typically the
maximum transmission frame size for most serial connections. Ethernet networks tend to
use serial devices to access the Internet.
Data/pad, which is also known as payload: The data is inserted in the data/pad or
payload. This location is where the IP header and data are placed if you are running IP
over Ethernet. This field contains Internetwork Packet Exchange (IPX) information if you
are running IPX/Sequenced Packet Exchange (SPX) protocol (Novell). The following
specific fields are contained within the data/pad section of an IEEE 803.2 frame:
Frame check sequence: This field contains the frame check sequence (FCS), which is
calculated by using a cyclic redundancy check (CRC). The FCS allows Ethernet to detect
errors in the Ethernet frame and reject the frame if the frame appears damaged.
65
All machines in the network are free to use the network whenever they choose if no one else
is transmitting. This setup is known as Multiple Access.
You are required to have a means of ensuring that when two machines start to transmit data
simultaneously, the resultant corrupted data is discarded. Also, retransmissions must be
generated at differing time intervals. This assurance is known as Collision Detection.
Figure 4-2 shows a bus Ethernet network.
In Figure 4-2, assume that machine 2 wants to send a message to machine 4, but first it
listens to make sure that no one else is using the network.
If the path is all clear, machine 2 starts to transmit its data on to the network. Each packet of
data contains the destination address, the sender address, and the data to be transmitted.
The signal moves down the cable, and the signal is received by every machine in the network.
But, because the signal it is only addressed to machine 4, the other machines ignore the
signal. Machine 4 then sends a message back to machine 2 to acknowledge the receipt of the
data.
But what happens when two machines try to transmit at the same time? A collision occurs,
and each machine has to back off for a random period before it tries to transmit again.
Figure 4-3 shows what happens when two machines transmit simultaneously.
Figure 4-3 Machine 2 and machine 5 are both trying to transmit simultaneously
The resulting collision in Figure 4-3 destroys both signals. Each machine knows that this
result happened because they do not hear their own transmission within a certain period.
This time period is the propagation delay. The propagation delay is equivalent to the time that
it takes for a signal to travel to the furthest part of the network and back again.
Both of the machines then wait for a random period before they try to transmit again. On small
networks, this process all happens so quickly that it is virtually unnoticeable. However, as
more machines are added to a network, the number of collisions rises dramatically and
eventually results in slow network response. The exact number of machines that a single
Ethernet segment can handle depends on the applications that are used, but the general
consideration is that 40 - 70 users are the limit before network speed is compromised.
66
Figure 4-4 shows two scenarios: hub and switch. The hub is where all of the machines are
interconnected so that only one machine at a time can use the media. In the switch network,
more than one machine can use the media at a time.
An Ethernet hub changes the topology from a bus to a star-wired bus. As an example,
assume again that machine 1 is transmitting data to machine 4. But this time, the signal
travels in and out of the hub to each of the other machines.
Collisions can still occur but hubs have the advantage of centralized wiring, and they can
automatically bypass any ports that are disconnected or have a cabling fault. This centralized
wiring makes the network much more fault tolerant than a coax-based system where
disconnecting a single connection shuts down the whole network.
With a switch, machines can transmit simultaneously. Each switch reads the destination
addresses and switches the signals directly to the recipients without broadcasting to all of the
machines in the network.
This point-to-point switching alleviates the problems that are associated with collisions and
considerably improves the network speed.
67
These figures go against the traditional use of binary prefixes for memory size. These decimal
prefixes are established in data communications.
Table 4-1 lists the technology rates and the medium.
Table 4-1 Technology rates and medium
Technology
Media
Fast Ethernet
(100BASE-X)
100 Mbps
12.5 MBps
UTP Cat 5
Gigabit Ethernet
(1000BASE-X)
1000 Mbps
125 MBps
10 Gigabit Ethernet
(10GBASE-X)
10000 Mbps
1250 MBps
4.1.5 10 GbE
From its origin more than 25 years ago, Ethernet evolved to meet the increasing demands of
packet-based networks. Ethernet provides the benefits of proven low implementation cost,
reliability, and relative simplicity of installation and maintenance. Because of these benefits,
the popularity of Ethernet grew to the point that nearly all of the traffic on the Internet
originates or terminates with an Ethernet connection. Furthermore, as the demand for
ever-faster network speeds increased, Ethernet adapted to handle these higher speeds and
the surges in volume demand that accompany them.
The IEEE 802.3ae 2002 (the 10 Gigabit Ethernet (10 GbE) standard) differs in certain
respects from earlier Ethernet standards in that it operates only in full-duplex mode.
(Collision-detection protocols are unnecessary.)
Ethernet can now progress to 10 gigabits per second while it retains its critical Ethernet
properties, such as the packet format. The current capabilities are easily transferable to the
new standard.
The 10 Gigabit Ethernet technology continues the evolution of Ethernet in terms of speed and
distance, while it retains the same Ethernet architecture that is used in other Ethernet
specifications. However, one key exception exists. Because 10 Gigabit Ethernet is a
full-duplex-only technology, it does not need the CSMA/CD protocol that is used in other
Ethernet technologies. In every other respect, 10 Gigabit Ethernet matches the original
Ethernet model.
68
Copper is the default standard for transmitting data between devices because of its low cost,
easy installation, and flexibility. It also possesses distinct shortcomings. Copper is best when
used in short lengths, typically 100 meters (328 feet) or less. When copper is employed over
long distances, electromagnetic signal characteristics hinder performance. In addition,
bundling copper cabling can cause interference, which makes it difficult to employ copper
cabling as a comprehensive backbone. For these reasons, copper cabling is the principal
data carrying technique for communication among personal computers and LANs, but not
campus or long-distance transmission.
Conversely, fiber cabling is typically used for remote campus connectivity, crowded
telecommunications closets, long-distance communications, and environments that need
protection from interference. An example of this environment is a manufacturing area.
Because fiber cabling is reliable and less susceptible to attenuation, fiber cabling is optimum
for sending data beyond 100 meters (328 ft).
However, fiber costs more than copper. Therefore, the use of fiber cabling is typically limited
to those applications that demand it.
As a result, most organizations use a combination of copper and fiber cabling. As these
companies transition to 10 GbE functionality, they must have a solid understanding of the
various cabling technologies. Companies must also have a sound migration strategy to
ensure that their cabling infrastructure will support their network infrastructure both today and
in the future.
The IEEE 802.3 Higher Speed Study Group formed in 1998, and the development of 10GigE
began the following year. By 2002, the 10GigE standard was first published as IEEE Std
802.3ae-2002. This standard defines a normal data rate of 10 Gigabits, making it 10 times
faster than the Gigabit Ethernet.
Subsequent standard updates ensued in relation to the first 10GigE version that was
published in 2002. The IEEE 802.3ae-2002 fiber and the 802.3ak-2004 in 2004 were later
consolidated into IEEE 802.3-2005 in 2005. In 2006, 802.3an-2006, which is a
10 Gigabit Base-T copper twisted pair, and an enhanced version with fiber-LRM PMD
followed, which were known as 802.3aq-2006. Finally, in 2007, the 802.3ap-2007 with copper
backplane evolved.
As a result of these standards, two major types of 10 Gigabit Ethernet cabling, fiber and
copper, are available.
The following standards apply to the 10 Gigabit Ethernet fiber cabling:
10GBASE-LX4: This standard supports ranges of 240 meters - 300 meters (790 ft - 980 ft)
over traditional multi-mode cabling. This range is achieved by using four separate laser
sources that operate at 3.125 Gbps in the range of 1300 nm on unique wavelengths. The
10GBASE-LX4 standard also supports 10 kilometers (6.2 miles) over
System Management Facilities (SMF).
10GBASE-SR: Over obsolete 62.5 micron multi-mode fiber cabling (OM1), this standard
has a maximum range of 26 meters - 82 meters (85 ft - 269 ft), depending on the cable
type. Over standard 50 m 2000 MHzkm OM3 multi-mode fiber (MMF), this standard has
a maximum range of 300 meters (980 ft).
10GBASE-LR: This standard has a specified reach of 10 kilometers (6.2 miles), but
10GBASE-LR optical modules can often manage distances of up to 25 kilometers
(16 miles) with no data loss.
69
70
71
Figure 4-6 shows two methods for maintaining isolation of VLAN traffic between switches.
The first method uses a single link for each VLAN. This method does not scale well because
it uses many ports in networks that have multiple VLANs and multiple switches. Also, this
method does not use link capacity efficiently when traffic in the VLANs is not uniform.
The second method is VLAN tagging over a single link in which each frame in tagged with its
VLAN ID. This method is highly scalable because only a single link is required to provide
connectivity to many VLANs. This configuration provides for better utilization of the link
capacity when VLAN traffic is not uniform.
The protocol for VLAN tagging of frames in a LAN environment is defined by the
72
Tagged frames
FCS - 4
Payload
46 - 1500 (Std)
up to 9K
2 (Jumbo)
1
Type - 2
Tag - 4
SA - 6
DA - 6
Type
0x8100
Priority - 3
CFI - 1
SFD - 1
Preamble - 7
The IEEE 802.1p/q standard provides a methodology for information, such as VLAN
membership and priority, that is added to the frame (Figure 4-7).
VLAN ID
12 bits
73
Variations in the meaning of trunk: The term trunk is used to express different ideas in
the networking industry. When you use this term, remember that other individuals might
use the term in a different manner. Trunk can mean that a port is operating in multiple
VLAN mode or it can mean a link aggregated port.
74
Link aggregation provides greater bandwidth between the devices at each end of the
aggregated link. Another advantage of link aggregation is increased availability because the
aggregated link is composed of multiple member links. If one member link fails, the
aggregated link continues to carry traffic over the remaining member links.
Each of the devices that is interconnected by the aggregated link uses a hashing algorithm to
determine on which of the member links the frames will be transmitted. The hashing algorithm
might use various information in the frame to make the decision. This algorithm might include
a source MAC, destination MAC, source IP, destination IP, and more. It might also include a
combination of these values.
75
All leaf devices calculate the best path to the root device. The devices place their ports in
blocking or forwarding states that are based on the best path to the root. The resulting tree
topology provides a single active Layer 2 data path between any two end stations.
76
Loop protection
Loop protection increases the efficiency of STP, RSTP, VSTP, and MSTP by preventing ports
from moving into a forwarding state that might result in a loop opening in the network.
A blocking interface can transition to a forwarding state in error if the interface stops receiving
BPDUs from its designated port on the segment. This transition error can occur when a
hardware error is on the switch or a software configuration error is between the switch and its
neighbor.
When loop protection is enabled, the spanning tree topology detects root ports and blocked
ports and ensures that both root ports and blocked ports keep receiving BPDUs. If a loop
protection-enabled interface stops receiving BPDUs from its designated port, it reacts as it
might react to a problem with the physical connection on this interface. It does not transition
the interface to a forwarding state, but instead it transitions the interface to a loop-inconsistent
state. The interface recovers and then transitions back to the spanning-tree blocking state as
soon as it receives a BPDU.
You must enable loop protection on all switch interfaces that have a chance of becoming root
or designated ports. Loop protection is the most effective when it is enabled on the entire
switched network. When you enable loop protection, you must configure at least one action
(alarm, block, or both).
An interface can be configured for either loop protection or root protection, but not for both.
Root protection
Root protection increases the stability and security of STP, RSTP, VSTP, and MSTP by
limiting the ports that can be elected as root ports. A root port that is elected through the
regular process has the possibility of being wrongly elected. A user bridge application that is
running on a personal computer can also generate BPDUs and interfere with root port
election. With root protection, network administrators can manually enforce the root bridge
placement in the network.
77
Root protection is enabled on interfaces that must not receive superior BPDUs from the root
bridge and must not be elected as the root port. These interfaces become designated ports
and are typically on an administrative boundary. If the bridge receives superior STP BPDUs
on a port that enabled root protection, that port transitions to a root-prevented STP state
(inconsistency state), and the interface is blocked. This blocking prevents a bridge that must
not be the root bridge from being elected the root bridge. After the bridge stops receiving
superior STP BPDUs on the interface with root protection, the interface returns to a listening
state. This state is followed by a learning state and ultimately back to a forwarding state.
Recovery back to the forwarding state is automatic.
When root protection is enabled on an interface, it is enabled for all of the STP instances on
that interface. The interface is blocked only for instances for which it receives superior
BPDUs. Otherwise, it participates in the spanning tree topology. An interface can be
configured for either root protection or loop protection, but not for both.
4.1.12 Link Layer Discovery Protocol Type Length Values (LLDP TLVs)
The basic TLVs include the following information:
Chassis identifier: The MAC address that is associated with the local system.
Port identifier: The port identification for the specified port in the local system.
Port description: The user-configured port description. This description can be a maximum
of 256 characters.
System name: The user-configured name of the local system. The system name can be a
maximum of 256 characters.
78
System description: The system description contains information about the software and
the current image that are running on the system. This information is not configurable, but
it is taken from the software.
System capabilities: The primary function that is performed by the system. The
capabilities that are supported by the system, for example, bridge or router. This
information is not configurable, but it is based on the model of the product.
Management address: The IP management address of the local system.
Additional 802.3 TLVs include the following details:
Power by way of medium dependent interface (MDI): A TLV that advertises MDI power
support, a Power Sourcing Equipment (PSE) power pair, and power class information.
MAC/PHY configuration status: A TLV that advertises information about the physical
interface, such as auto-negotiation status, support, and multistation access unit (MAU)
type. The information is not configurable, but it is based on the physical interface structure.
Link aggregation: A TLV that advertises whether the port is aggregated and its aggregated
port ID.
Maximum frame size: A TLV that advertises the maximum transmission unit (MTU) of the
interface that is sending LLDP frames.
Port VLAN: A TLV that advertises the VLAN name that is configured on the interface.
LLDP-MED provides the following TLVs:
LLDP MED capabilities: A TLV that advertises the primary function of the port. The
capability values range 0 - 15. The device class values range 0 - 255:
0: Capabilities
1: Network policy
2: Location identification
3: Extended power by way of MDI-PSE
4: Inventory
5 - 15: Reserved
Network policy: A TLV that advertises the port VLAN configuration and associated Layer 2
and Layer 3 attributes. The following attributes are included:
Policy identifier
Application types, such as voice or streaming video
802.1Q VLAN tagging
802.1p priority bits
Diffserv code points
Endpoint location: A TLV that advertises the physical location of the endpoint.
Extended power by way of MDI: A TLV that advertises the power type, power source,
power priority, and power value of the port. It is the responsibility of the PSE device
(network connectivity device) to advertise the power priority on a port.
79
People do not want to make a large financial investment without knowing that they will get a
return. The appeal of these protocols is that they immediately provide benefits. Because
these protocols are standards-based protocols, they allow the use of both the existing TCP/IP
and Fibre Channel Protocol (FCP) infrastructure, they support existing Fibre Channel
devices, and they enable the simplification of the infrastructure by removing any SAN islands.
4.2.4 Tunneling
Tunneling is a technique that allows one network to send its data through the connection of
another network. Tunneling works by encapsulating a network protocol within packets that are
carried by the second network. For example, in a Fibre Channel over Internet Protocol (FCIP)
solution, Fibre Channel packets can be encapsulated inside IP packets. Tunneling raises
issues of packet size, compression, out-of-order packet delivery, and congestion control.
80
IP
Hdr
FC
Hdr
SCSI data
CRC EOF
CRC
IP datagram
Figure 4-10 FCIP encapsulates the Fibre Channel frame into IP packets
81
The major advantage of FCIP is that it overcomes the distance limitations of basic Fibre
Channel. It also enables geographically distributed devices to be linked by using the existing
IP infrastructure, while it keeps the fabric services intact.
The architecture of FCIP is outlined in the Internet Engineering Task Force (IETF) Request for
Comments (RFC) 3821, Fibre Channel over TCP/IP (FCIP), which is available at this
website:
http://ietf.org/rfc/rfc3821.txt
Because FCIP simply tunnels Fibre Channel, creating an FCIP link is similar to creating an
ISL. And, the two fabrics at either end are merged into a single fabric. This merger creates
issues in situations where you do not want to merge the two fabrics for business reasons, or
where the link connection is prone to occasional fluctuations.
Many corporate IP links are robust, but it can be difficult to be sure because traditional
IP-based applications tend to be retry-tolerant. Fibre Channel fabric services are not as
retry-tolerant. Each time the link disappears or reappears, the switches renegotiate and the
fabric is reconfigured.
By combining FCIP with Fibre Channel-to-Fibre Channel (FC-FC) routing, the two fabrics can
be left unmerged, each with its own separate Fibre Channel services.
82
A myth exists that iFCP does not use encapsulation. In fact, iFCP encapsulates the Fibre
Channel packet in much the same way that FCIP encapsulates the Fibre Channel packet. In
addition, iFCP maps the Fibre Channel header to the IP header and a TCP session
(Figure 4-11).
FC
Hdr
SCSI data
CRC EOF
d
pe
ap
M
IP
TCP
Hdr
CRC
IP datagram
Figure 4-11 iFCP encapsulation and header mapping
83
A session between a iSCSI initiator and an iSCSI target is defined by a session ID. This
session ID is a combination of an initiator part (ISID) and a target part (Target Portal Group
Tag).
The iSCSI transfer direction is defined in relation to the initiator. Outbound or outgoing
transfers are transfers from an initiator to a target. Inbound or incoming transfers are transfers
from a target to an initiator.
For performance reasons, iSCSI allows a phase-collapse. A command and its associated
data might be shipped together from initiator to target, and data and responses might be
shipped together from targets.
An iSCSI name specifies a logical initiator or target. It is not tied to a port or hardware
adapter. When multiple NICs are used, they generally all present the same iSCSI initiator
name to the targets because they are paths to the same SCSI layer. In most operating
systems, the named entity is the operating system image.
The architecture of iSCSI is outlined in IETF RFC 3720, Internet Small Computer Systems
Interface (iSCSI), at this website:
http://www.ietf.org/rfc/rfc3720.txt
Figure 4-12 shows the format of the iSCSI packet.
Testing on iSCSI latency shows a difference of up to 1 ms of additional latency for each disk
I/O as compared to Fibre Channel. This delay does not include factors, such as trying to
perform iSCSI I/O over a shared, congested, or long-distance IP network, all of which might
be tempting for certain clients. iSCSI generally uses a shared 1 Gbps network.
84
85
Certain vendors tweak the back-off and recovery algorithms. For example, the tweak causes
the send rate to drop by 12.5% each time that congestion is encountered. And the algorithm
is tweaked so that the network can recover rapidly to the full sending rate by doubling each
time until the full rate is regained.
Other vendors take a simpler approach to achieve a similar outcome.
If you are sharing your IP link between storage and other IP applications, either of these
storage-friendly congestion controls might affect your other applications.
For more information about the specification for TCP congestion control, see this website:
http://www.ietf.org/rfc/rfc2581.txt
86
For smaller frames, such as the Fibre Channel acknowledgment frames, the time in transit is
much shorter. The minimum possible Fibre Channel frame is one with no payload. With FCIP
encapsulation, the minimum size of a packet with only the headers is 90 bytes, or 720 bits.
Table 4-2 lists the transmission times of this FCIP packet over various common wide area
network (WAN) link speeds.
Table 4-2 FCIP packet transmission times over different WAN links
Link type
Link speed
Large packet
Small packet
Gigabit Ethernet
1250 Mbps
14 s
0.6 s
OC-12
622.08 Mbps
28 s
1.2 s
OC-3
155.52 Mbps
113 s
4.7 s
T3
44.736 Mbps
394 s
16.5 s
E1
2.048 Mbps
8600 s
359 s
T1
1.544 Mbps
11 400 s
477 s
If we cannot use jumbo frames, each large Fibre Channel frame must be divided into two
Ethernet packets. This requirement doubles the amount of TCP, IP, and Ethernet MAC
overhead for the data transfer.
Normally, each Fibre Channel operation transfers data in only one direction. The frames that
move in the other direction are close to the minimum size.
87
In larger fabrics with many switches and separate business units, for example, in a shared
services-hosting environment, separation and routing are valuable. These features are
beneficial in creating many simple fabrics, rather than a few more complex fabrics.
88
Chapter 5.
89
90
91
We describe FC-AL in more depth in 5.4, Fibre Channel Arbitrated Loop protocols on
page 114.
92
This configuration is one of the major reasons why arbitrated loop is considered a historical
SAN topology. A switched fabric is typically referred to as a fabric.
In terms of switch interconnections, the switched SAN topologies can be classified as the
following types:
Single switch topology
Cascaded and ring topology
Mesh topology
93
Figure 5-5 indicates a single switch topology with all of the devices connected to same switch.
Even in a ring topology, the switches connect in a queue fashion, but the ring topology forms
a closed ring with an additional ISL (Figure 5-7).
94
In terms of a tiered approach, the switched fabric can be further classified with the following
topologies:
Core-edge topology
Edge-core-edge topology
95
96
97
G_port: This generic port type can operate as either an expansion port (E_port) or an
F_port. A port is defined as a G_port after it connects, but it did not receive a response to
loop initialization, or it did not yet complete the link initialization procedure with the
adjacent Fibre Channel device.
L_port: This loop port type is a loop-capable node or switch port.
U_port: This type is a universal port: a more generic switch port than a G_port. It can
operate as either an E_port, F_port, or FL_port. A port is defined as a U_port when it is
not connected or it did not yet assume a specific function in the fabric.
N_port: This port type is a node port that is not loop capable. It is a host end port that is
used to connect to the fabric switch.
NL_port: This port type is a node port that is loop capable. It is used to connect an
equipment port to the fabric in a loop configuration through an L_port or FL_port.
Figure 5-11 shows the common port types of the switch and nodes.
98
VE_Port: A virtual E_Port is a port that emulates an E_Port over a Fibre Channel over
Internet Protocol (FCIP) link. VE_Port connectivity is supported over point-to-point links.
VEX_Port: VEX_Ports are routed VE_Ports, just as Ex_Ports are routed E_Ports.
VE_Ports and VEX_Ports behave and function in the same way.
TE_port: The TE_port provides standard E_port functions and it also allows for the routing
of multiple virtual SANs (VSANs). This capability is accomplished by modifying the
standard Fibre Channel frame (VSAN tagging) on ingress and egress of the VSAN
environment. It is also known as a Trunking E_port.
Figure 5-12 shows a fabric with expansion ports.
99
Figure 5-13 shows the various test options. You can perform long-distance cable checks also
with D_port diagnostic capabilities.
100
Figure 5-14 shows the Fibre Channel port types that are specific to Cisco.
5.3 Addressing
All devices in a Fibre Channel environment have an identity. The way that the identity is
assigned and used depends on the format of the Fibre Channel fabric. For example,
addressing is performed differently in an arbitrated loop and in a fabric.
101
Figure 5-15 shows both the old and new WWN formats.
The new addressing scheme starts with a hex 5 or 6 in the first half-byte, which is followed by
the vendor identifier in the next 3 bytes. The vendor-specific information is then contained in
the following fields. Both of these formats are currently in use and they depend on the
hardware manufacturer standards to follow either of the formats. However, the vendor ID and
company ID are assigned uniquely by the IEEE standards, and each vendor and its identifier
are in the following text file:
http://standards.ieee.org/develop/regauth/oui/oui.txt
A worldwide node name (WWNN) is a globally unique 64-bit identifier that is assigned to each
Fibre Channel node or device. For servers and hosts, the WWNN is unique for each host bus
adapter (HBA). For a server with two HBAs, each HBA has a WWNN (two WWNNs total for
the server). For a SAN switch, the WWNN is common for the chassis. For storage, the
WWNN is common for each controller unit of midrange storage. And, in high-end enterprise
storage, the WWNN is unique for the entire array.
A worldwide port number (WWPN) is a unique identifier for each FC port of any Fibre
Channel device. For a server, we have a WWPN for each port of the HBA. For a switch, the
WWPN is available for each port in the chassis; and for storage, each host port has an
individual WWPN.
102
Figure 5-16 Server worldwide node name and worldwide port name
103
Storage area network worldwide node name and worldwide port name
Figure 5-17 shows that the WWNN is for the entire SAN switch chassis and the WWPN is for
each FC port in the SAN switch chassis.
Fabric-assigned PWWNs: The new 16 Gbps b-type switches with Brocade Fabric OS
(FOS) 7.0 can also have a virtual WWPN that is defined by switches that are called
fabric-assigned PWWNs (FAPWWNs). These FAPWWNs can be used for pre-configuring
zoning before the physical servers are connected. This feature helps to simplify and
accelerate server deployment and improve operational efficiency by avoiding the wait time
for setting up physical connectivity. This feature also requires that servers use Brocade
HBAs/adapters with an HBA driver version 3.0.0.0 or higher, which can be configured to
use FAPWWN.
Figure 5-17 SAN switch worldwide node name and worldwide port name
104
Figure 5-18 Storage worldwide node name and worldwide port name
Worldwide node name (WWNN): The IBM virtualization storage systems use WWNN
differently. For example, each node in an IBM SAN Volume Controller or the
IBM Storwize V7000 has an individual and unique WWNN.
For the IBM DS8000, each Storage Facility Image has a unique individual WWNN.
5.3.2 Tape device worldwide node name and worldwide port name
For tape devices, each drive inside the tape library has an individual WWPN and WWNN.
Figure 5-19 shows that multiple drive libraries have an individual WWNN and WWPN for each
drive.
Figure 5-19 Tape device worldwide node name and worldwide port name
105
106
Domain
Bits 23
Port
Area
16 15
08 07
00
AL_PA
The following functions provide the significance of several of the bits that make up the port
address:
Domain
The most significant byte of the port address is the domain. This byte is the address of the
switch. A domain ID is a unique number that identifies the switch or director to a fabric. It
can be either static or dynamic. Static (insistent) domain IDs are a requirement for Fibre
Channel connection (FICON). Each manufacturer has a range of numbers and a
maximum number of domain IDs that can be used in a fabric.
One byte allows up to 256 possible addresses. Because many of these addresses are
reserved, such as the address for broadcast, only 239 addresses are available. This
number means that you can theoretically have as many as 239 switches in your SAN
environment. The domain number allows each switch to have a unique identifier if you
have multiple interconnected switches in your environment.
Area
The area field provides 256 addresses. This part of the address is used to identify the
individual ports. Therefore, to have more than 256 ports in one switch in a director class of
switches, you must follow the shared area addressing.
Port
The final part of the address provides 256 addresses for identifying attached N_ports and
NL_ports.
A simple calculation is used to arrive at the number of available addresses:
Domain x area x ports
This calculation means that 239 x 256 x 256 = 15,663,104 addresses are available.
Depending on the fabric topology, the fabric addressing format of the device differs.
In a fabric topology, devices have an addressing format type of DDAA00. For example, the
address 020300 indicates that the device belongs to the switch with domain ID 02. This
switch is connected to port 03 and the ALPA address is 00, which indicates that this device is
not a loop fabric device. That is, it is a switched fabric device. For any switched fabric device,
the ALPA ID is always 00.
107
Native mode is used in traditional switches that support a maximum of 16 ports. This number
is used because in native mode, the fabric addressing format that is used is DD1A00. The
area part of the fabric address always has a prefix of 1. Therefore, it supports a port count
from hexadecimal 10 to 1F (a maximum of 16 ports).
Core PID mode is used to support a maximum of 256 ports for each domain or switch. This
number is used because in core PID mode, the area part of the fabric address supports
addresses from hexadecimal 00 to FF (a maximum of 256 ports). The fabric addressing
format that is used for this mode is DDAA00.
108
Figure 5-21 shows the native and core PID modes with the example FC address of two
devices.
Shared addressing mode is used when more than 256 ports are used in the same domain or
switch. This mode is used in directors with high port density. The port addressing in these
directors uses the same area numbers for two ports by having the third byte of the
FC address (node addresses) as 80 for higher port numbers. By using the area ID more than
one time, this mode enables more than 256 ports to exist in a single domain.
Figure 5-22 shows port 24. Port 25 shares the area ID with port 32, and port 33 of the FC4-48
port.
109
110
The FICON CU port also performs a fabric login to obtain its 24-bit FC port address.
Figure 5-23 shows how the FC-FS 24-bit FC port address identifier is divided into three fields.
Figure 5-23 shows the FC-FS 24-bit port address and the definition of usage of that 24-bit
address in an IBM z Systems server and 9672 G5/G6 environment. Only the 8 bits that make
up the FC port address are defined for the z Systems server and 9672 G5/G6 to access a
FICON CU.
The FICON channel in FICON native (FC) mode that works with a switched point-to-point FC
topology (single switch) provides the other 2 bytes that make up the 3-byte FC port address of
the CU to be accessed.
The z Systems and 9672 G5/G6 processors, when they work with a switched point-to-point
topology, require that the Domain and the AL_port (Arbitrated Loop) field values are the
same for all of the FC F_ports in the switch. Only the area field value differs for each switch
F_port.
For the z Systems server and 9672 G5/G6, the area field is referred to as the port address
field of the F_port. This field is only a 1-byte value. When the access to a CU that is attached
to this port is defined, by using the z Systems HCD or IOCP, the port address is referred to as
the link address.
111
The 8 bits for the domain address and the 8-bit constant field are provided from the Fabric
Login initialization result (Figure 5-24). The 8 bits and the 1 byte for the port address (1-byte
link address) are provided from the z Systems or 9672 G5/G6 CU link definition (by using
HCD and IOCP).
112
Figure 5-25 shows that the FC-FS 24-bit FC port address identifier is divided into three fields:
Domain
Area
AL (Port)
Figure 5-25 shows the FC-FS 24-bit port address and the definition usage of that 24-bit
address in a z Systems environment. The 16 bits that make up the FC port address must be
defined for the z Systems to access a FICON CU in a cascaded environment. The FICON
channel in the FICON native (FC) mode that works with a two-switch cascaded FC topology
provides the remaining byte that makes up the full 3-byte FC port address of the CU to be
accessed.
It is required that the Domain, Switch @, AL_Port, and the Arbitrated Loop field values are the
same for all of the FC F_ports in the switch. Only the Area field value differs for each switch
F_port.
The z Systems Domain and Area fields are referred to as the port address field of the F_port.
This field is a 2-byte value. When access is defined to a CU that is attached to this port (by
using the z Systems HCD or IOCP), the port address is referred to as the link address.
113
The 8 bits for the constant field are provided from the Fabric Login initialization result
(Figure 5-26). The 16 bits for the port address and 2-byte link address are provided from the z
Systems CU link definition by using HCD and IOCP.
As a footnote, FCP connectivity is device-centric and defined in the fabric by using the WWPN
of the devices that are allowed to communicate. When an FCP device attaches to the fabric, it
queries the name server for the list of devices that it is allowed to form connections with (that
is, the zoning information). FICON devices do not query the name server for accessible
devices because the allowable port and device relationships are defined on the host.
Therefore, the zoning and name server information does not need to be retrieved.
114
115
116
When the communication parameters and identities of other devices are discovered, they are
able to establish logical sessions between devices (initiator and targets).
117
118
Management services
Time services
Simple name server
Login services
Registered State Change Notification (RSCN)
These services are implemented by switches and directors that participate in the SAN.
Generally, the services are distributed across all of the devices, and a node can use the
switching device to which it is connected.
All of these services are addressed by FC-2 frames, and they are accessed by well-known
addresses.
119
120
121
5.8 Zoning
Zoning allows for finer segmentation of the switched fabric. Zoning can be used to instigate a
barrier between different environments. Only the members of the same zone can
communicate within that zone; all other attempts from outside are rejected.
For example, you might want to separate a Microsoft Windows environment from a UNIX
environment because of the manner in which Windows attempts to claim all available storage
for itself. Because not all storage devices can protect their resources from any host that seeks
available resources, a preferred practice is to protect the environment in another manner. We
show an example of zoning in Figure 5-30 where we separate AIX from Windows and create
Zone 1 and Zone 2. This diagram also shows how a device can be in more than one zone.
Looking at zoning in this way, consider zoning as a security feature and not merely for
separating environments. Zoning can also be used for test and maintenance. For example,
not many enterprises mix their test and maintenance environments with their production
environment. Within a fabric, you can easily separate your test environment from your
production bandwidth allocation on the same fabric by using zoning.
122
B
A&B
Zoning also introduces the flexibility to manage a switched fabric to meet the objectives of
separate user groups.
Zoning can be implemented in the following ways:
Hardware zoning
Software zoning
These forms of zoning are different, but they are not necessarily mutually exclusive.
Depending on the particular manufacturer of the SAN hardware, hardware zones and
software zones can overlap. Although this ability adds to the flexibility, it can complicate the
solution, increasing the need for good management software and SAN documentation.
123
Figure 5-32 shows an example of zoning that is based on the switch port numbers.
Server A
Server B
ESS A
Server C
Storage
Area
Network
Tape A
ESS B
Tape
TapeB
124
Figure 5-33 shows an example of hardware zoning. This example illustrates another way to
consider the hardware zoning as an array of connections.
In Figure 5-33, device A can access only storage device A through connection A. Device
B can access only storage device B through connection B.
In a hardware-enforced zone, switch hardware, usually at the application-specific integrated
circuit (ASIC) level, ensures that no data is transferred between unauthorized zone members.
However, devices can transfer data between ports within the same zone. Therefore, hard
zoning provides the highest level of security. The availability of hardware-enforced zoning and
the methods to create hardware-enforced zones depend on the switch hardware.
One disadvantage of hardware zoning is that devices must be connected to a specific port,
and the whole zoning configuration can become unusable when the device is connected to a
different port. In cases where the device connections are not permanent, the use of software
zoning is likely to simplify your configuration.
The advantage of hardware zoning is that it can be implemented into a routing engine by
filtering. As a result, this type of zoning has a low effect on the performance of the routing
process.
If possible, the designer can include several unused ports in a hardware zone. Therefore, if a
particular port fails, perhaps because of a gigabit interface converter (GBIC) or transceiver
problem, the cable can be moved to a different port in the same zone. Therefore, the zone
does not need to be reconfigured.
125
126
Figure 5-34 shows an example of WWN-based zoning. In this example, symbolic names are
defined for each WWN in the SAN to implement the same zoning requirements that were
shown in Figure 5-32 on page 124 for port zoning:
Zone_1 contains the aliases alex, ben, and sam, and it is restricted to only these devices.
Zone_2 contains the aliases robyn and ellen, and it is restricted to only these devices.
Zone_3 contains the aliases matthew, max, and ellen, and it is restricted to only these
devices.
Server A
alex
ESS A
Server B
Server C
robyn
matthew
Storage
Area
Network
ben
max ESS B
sam
ellen
Tape A
Tape
TapeB
Zone
Name
Zone_1
Zone_2
Zone_3
Aliases
alex, ben, sam
robyn, ellen
matthew, max, ellen
Alias
WWPN
robyn
alex
matthew
ben
sam
max
ellen
50:05:76:ab:cd:22:03:65
50:05:76:ab:cd:12:06:92
50:05:76:ab:cd:24:05:94
50:05:76:ab:cd:20:09:91
50:05:76:ab:cd:23:05:93
50:05:76:ab:cd:02:05:94
50:05:76:ab:cd:20:08:90
127
Many switch vendors offer hardware-enforced WWN zoning, which can prevent this security
exposure. Hardware-enforced zoning uses hardware mechanisms to restrict access rather
than relying on the servers to follow the Fibre Channel Protocols.
Software zoning: When a device logs in to a software-enforced zone, it queries the name
server for devices within the fabric. If zoning is in effect, only the devices in the same zone
or zones are returned. Other devices are hidden from the name server query reply. When
you use software-enforced zones, the switch does not control data transfer, and no
guarantee exists for data that is transferred from unauthorized zone members. Use
software zoning where flexibility and security are ensured by the cooperating hosts.
Frame filtering
Zoning is a fabric management service that can be used to create logical subsets of devices
within a SAN. This service can also enable the partitioning of resources for management and
access control. Frame filtering is another feature that enables devices to provide zoning
functions with finer granularity. Frame filtering can be used to set up port-level zoning, WWN
zoning, device-level zoning, protocol-level zoning, and logical unit number (LUN)-level
zoning. Frame filtering is commonly performed by an application-specific integrated circuit
(ASIC). Use this configuration, after the filter is set up, to achieve the complicated function of
zoning and filtering at wire speed.
128
Chapter 6.
129
Servers
Application
Laptops
Monitoring
Content
Collaboration
Communication
Finance
Platform
Identity
Object Storage
Desktops
Runtime
Queue
Database
Infrastructure
Compute
Phones
Network
Block Storage
Tablets
Cloud Computing
Figure 6-1 Cloud computing overview
Cloud computing provides computation, software, data access, and storage services that do
not require user knowledge of the physical location and configuration of the system that
delivers the services. Parallels to this concept can be drawn with the electricity grid, wherein
users use power without needing to understand the component devices or the infrastructure
that is required to provide the service.
Cloud computing describes a new consumption and delivery model for IT services, and it
typically involves provisioning of dynamically scalable and virtualized resources. The cloud
introduces three key concepts: cost savings, service reliability, and infrastructure flexibility.
To cater to the increasing, on-demand needs of business, IT services and infrastructures are
moving rapidly toward a flexible utility and consumer model by adopting new technologies.
One of these technologies is virtualization. Cloud computing is an example of a virtual,
flexible delivery model. Inspired by consumer Internet services, cloud computing puts the
user in the drivers seat; that is, users can use Internet offerings and services by using this
self-service, on-demand model.
130
Cloud computing can potentially affect your business dramatically by providing the following
benefits:
From a technical perspective, cloud computing enables these capabilities, among others:
Abstraction of resources
Dynamic right-sizing
Rapid provisioning
Cloud Services
Cloud Infrastructure
Cloud Platform
SAN and Storage
Cloud Services
This layer is the service that is delivered to the client. It can be an application, a desktop, a
server, or disk storage space. The client does not need to know where or how their service is
running; they just use it.
131
Cloud Infrastructure
This layer can be difficult to visualize depending on the final delivered service. If the final
service is a chat application, the cloud infrastructure is the servers on which the chat
application is running. In the other case, if the final service is a virtualized server, the cloud
infrastructure is all of the other servers that are required to provide a server as a service to
the client. Examples of these types of servers include a domain name server (DNS), security
services, and management.
Cloud Platform
This layer consists of the selected platform to build the cloud. Many vendors exist, such as
IBM Smart Business Storage Cloud, VMware vSphere, Microsoft Hyper V, and Citrix Xen
Server, which are well-known cloud solutions in the market.
For this book, we describe Fibre Channel or FCoE for networking and compatible storage
devices.
132
133
From a storage perspective, IBM clients, based on their business requirements, can choose
to adopt either a public or private storage cloud. The following definitions describe these types
of storage clouds:
Public storage cloud
This type is designed for clients who do not want to own, manage, or maintain the storage
environment, therefore reducing their capital and operational expenditures for storage.
IBM dictates the choice of technology and cloud location, shared infrastructure with
variable monthly charges, dynamic physical capacity at the client level, and security
measures to isolate client data. The public storage cloud allows for variable billing options
and shared tenancy of the storage cloud, giving clients the flexibility to manage the use
and growth of their storage needs. This type is the industry-standard view of a storage
cloud offering and comparable to storage cloud offerings by other vendors.
Private storage cloud
With a private storage cloud, clients have the choice of technology and location on a
dedicated infrastructure with fixed monthly charges and a physical capacity that is
manageable by the client. Each application can use dynamic capacity by sharing the cloud
storage among multiple applications.
Private storage cloud solution technology and services from IBM address multiple areas of
functionality. For more information, see this website:
http://www.ibm.com/cloud-computing/us/en/
Hypervisor software is the secret that makes virtualization possible. This software sits
between the hardware and the operating system, and decouples the operating system and
applications from the hardware. The hypervisor assigns the amount of access that the
operating systems and applications have with the processor and other hardware resources,
such as memory and disk input/output (I/O).
In addition to using virtualization technology to partition one machine into several virtual
machines, you can also use virtualization solutions to combine multiple physical resources
into a single virtual resource. A good example of this solution is storage virtualization. This
type of virtualization is where multiple network storage resources are pooled into what is
displayed as a single storage device for easier and more efficient management of these
resources.
134
Other types of virtualization you might hear about include the following examples:
Network virtualization splits available bandwidth in a network into independent channels
that can be assigned to specific servers or devices.
Application virtualization separates applications from the hardware and the operating
system, putting them in a container that can be relocated without disrupting other systems.
Desktop virtualization enables a centralized server to deliver and manage individualized
desktops remotely. This type of virtualization gives users a full client experience. IT staff
can provision, manage, upgrade, and patch desktops virtually, instead of physically.
Virtualization was first introduced in the 1960s by IBM. It was designed to boost the utilization
of large, expensive mainframe systems by partitioning them into logical, separate virtual
machines to run multiple applications and processes at the same time. In the 1980s and
1990s, this centrally shared mainframe model gave way to a distributed, client/server
computing model, in which many low-cost x86 servers and desktops independently run
specific applications.
135
Server virtualization
Three popular approaches to server virtualization are available:
Virtual machine model
Paravirtual machine model
Virtualization at the operating system layer
136
Virtual machines (VMs) are based on the host/guest design. Each guest runs on a virtual
implementation of the hardware layer. This approach allows the guest operating system to run
without modifications. This way, the administrator can create guests that use different
operating systems. The guest has no knowledge of the host operating system because the
guest is unaware that it is not running on real hardware. This model, however, requires real
computing resources from the host so this model uses a hypervisor to coordinate instructions
to the CPU.
The paravirtual machine (PVM) model is also based on the host/guest design, and it uses a
virtual machine monitor (VMM). In the paravirtual machine model, however, the VMM actually
modifies the code of the guest operating system. This modification is called porting. Porting
supports the VMM so that the VMM can use privileged system calls sparingly. Paravirtual
machines also can run multiple operating systems. Xen and Unified Modeling Language
(UML) both use the paravirtual machine model.
Virtualization at the OS level works slightly differently. It is not based on the host/guest
design. In the OS level model, the host runs a single OS kernel as its core and exports the
operating system functionality to each of the guests. Guests must use the same operating
system as the host, although different distributions of the same system are allowed. This
distributed architecture eliminates system calls between layers, reducing CPU usage
overhead. This model also requires each partition to remain strictly isolated from its neighbors
so that a failure or security breach in one partition is unable to affect any of the other
partitions. In this model, common binary files and libraries on the same physical machine can
be shared, allowing an OS-level virtual server to host thousands of guests at the same time.
IBM AIX virtual I/O (VIO) and Solaris Zones both use OS-level virtualization.
Desktop Virtualization
Desktop virtualization is sometimes referred to as client virtualization. Desktop virtualization
is defined as a virtualization technology that is used to separate a computer desktop
environment from the physical computer. Desktop virtualization is considered a type of
client/server computing model because the virtualized desktop is stored on a centralized, or
remote, server and not on the physical machine that is virtualized.
Desktop virtualization virtualizes desktop computers. These virtual desktop environments are
served to users in the network. Users interact with a virtual desktop in the same way that
they access and use a physical desktop. Another benefit of desktop virtualization is that you
can log in remotely to access your desktop from any location.
One of the most popular uses of desktop virtualization is in the data center, where
personalized desktop images for each user are hosted on a data center server.
Also, options are available for using hosted virtual desktops, where the desktop virtualization
services are provided to a business through a third party. The service provider provides the
managed desktop configuration, security, and SAN.
Application Virtualization
Application virtualization is similar to desktop virtualization, where individual desktop
sessions (OS and applications) are virtualized and run from a centralized server. However,
application virtualization virtualizes the applications so that an application can either be run
from a centralized server or streamed from a central server and run in an isolated
environment on the desktop.
137
In the first type of application virtualization, the application image is loaded on to a central
server. When a user requests the application, the application is streamed to an isolated
environment on the users computer for execution. The application starts running shortly after
it gets sufficient data to start running, and because the application is isolated from other
applications, conflicts are less likely. The applications that can be downloaded can be
restricted based on the user ID, which is established by logging in to corporate directories,
such as Active Directory (AD) or Lightweight Directory Access Protocol (LDAP).
In the second type of application virtualization, the applications are loaded as an image in
remote servers and the applications are run (executed) in the servers. Only the on-screen
information that is required to be seen by the user is sent over the LAN. This type of
application virtualization is closer to desktop virtualization, but only the application is
virtualized instead of both the application and the operating system. The greatest advantage
of this type of application virtualization is that it does not matter what the underlying OS is in
the users computer because the applications are processed in the server. Another advantage
is the effectiveness of mobile devices, such as mobile phones and tablet computers, with less
processing power when the user runs applications that require significant processing
capabilities. These applications are processed in the powerful processors of the servers.
138
Tape virtualization
Tape virtualization refers to the virtualization of tapes and tape drives that use specialized
hardware and software. This type of virtualization can enhance backup and restore flexibility
and performance because disk devices are used in the virtualization process, rather than
tape media.
139
Logical fabric
When the fabric is formed with at least one logical switch, the fabric is called a logical fabric.
Two methods of fabric connectivity are available for logical fabrics:
A logical fabric is connected with a dedicated inter-switch link (ISL) to another switch or a
logical switch. Figure 6-7 shows a logical fabric that is formed between logical switches
through a dedicated ISL for logical switches.
Logical fabrics are connected by using a shared ISL, which is called an extended ISL
(XISL), from a base logical switch. In this case, the separate logical switch is configured as
a base switch. This separate logical switch is used only for XISL connectivity and not for
device connectivity. Figure 6-8 shows a logical fabric that is formed through the XISL in the
base switch.
Figure 6-8 Logical ISL formed through the XISL in the base switch
140
141
Figure 6-10 shows the VSAN across chassis with an enhanced inter-switch link (EISL) for
VSAN communication.
142
Figure 6-11 shows sharing a single HBA by multiple virtual nodes. In this case, the same HBA
is defined with multiple virtual worldwide node names (WWNNs) and worldwide port names
(WWPNs).
143
Figure 6-12 Blade server with FC switch module in the NPIV mode
With the NPIV mode, we can overcome the interoperability issues of merging external
switches that might come from separate vendors to the blade server switch module. Also,
management is easier because the blade switch module becomes a node in the fabric. And,
we can overcome the scalability limitations of many switch domains for a switch module in
blade servers.
144
Thin provisioning allows a server to see logical volume sizes that are larger than the physical
capacity that is dedicated to the volumes on the storage system. From the servers or
applications perspective, thinly provisioned volumes are displayed and function the same as
fully provisioned volumes. However, physical disk drive capacity is allocated only as needed
(on demand) for write activity to the volumes. Deallocated physical capacity is available for
use as needed by all volumes in a storage pool or even across an entire storage system.
Thin provisioning offers these advantages:
It allows higher storage systems utilization, which in turn leads to a reduction in the
amount of storage that you need, lowering your direct capital expenditure (CAPEX).
It lowers your operational expenditure (OPEX) because your storage occupies less data
center space and requires less electricity and cooling.
It postpones the need to buy more storage. And, as storage prices continue to drop over
time, when you require more capacity, the storage will likely cost less.
Capacity planning is simplified because you can manage a single pool of free storage.
Multiple applications or users can allocate storage from the same free pool, avoiding the
situation where certain volumes are capacity-constrained and other volumes have spare
capacity.
Your storage environment becomes more agile, and it is easier to react to change.
145
Storage utilization is measured by comparing the amount of physical capacity that is used for
data with the total amount of physical capacity that is allocated to a server. Historically,
utilization ratios were under 50%, indicating a large amount of allocated but unused physical
storage capacity. Often, the users and storage administrators are uncertain how much
capacity is needed. However, they must ensure that they do not run out of space, and they
also must allow for growth. As a result, users might request more capacity than they need and
storage administrators might allocate more capacity than is requested, resulting in a
significant over-allocation of storage capacity.
Thin provisioning increases storage utilization ratios by reducing the need to over-allocate
physical storage capacity to prevent out-of-space conditions. Large logical or virtual volume
sizes might be created and presented to applications without dedicating an equivalent
amount of physical capacity. Physical capacity can be allocated on demand as needed for
writing data. Deallocated physical capacity is available for multiple volumes in a storage pool
or across the entire storage system.
Thin provisioning also increases storage efficiency by reducing the need to resize volumes or
add volumes and restripe data as capacity requirements grow. Without thin provisioning, if an
application requires capacity beyond the capacity that is provided by its current set of
volumes, two options are available:
Existing volumes might be increased in size.
Additional volumes might be provisioned.
In many environments, these options are challenging because of the required steps and
potential disruption to make the volumes larger or additional volumes visible and optimized for
the application.
With thin provisioning, large virtual or logical volumes might be created and presented to
applications while the associated physical capacity grows only as needed, transparent to the
application.
Without thin provisioning, physical capacity was dedicated at the time of volume creation, and
storage systems typically did not display or report how much of the dedicated physical
capacity was used for data. As storage systems implemented thin provisioning, physical
allocation and usage became visible. Thin provisioning increases storage efficiency by
making it easy to see the amount of physical capacity that is needed and used because
physical space is not allocated until it is needed for data.
146
Data deduplication is a technology that reduces the amount of space that is required to store
data on disk. It achieves this space reduction by storing a single copy of data that is backed
up repetitively.
Data deduplication products read data while they look for duplicate data. Data deduplication
products break up data into elements and create a signature or identifier for each data
element. Then, they compare the data element signature to identify duplicate data. After they
identify duplicate data, they retain one copy of each element. They create pointers for the
duplicate items, and discard the duplicate items.
The effectiveness of data deduplication depends on many variables, including the rate of data
change, the number of backups, and the data retention period. For example, if you back up
the same incompressible data one time a week for six months, you save the first copy and you
do not save the next 24. This method provides a 25:1 data deduplication ratio. If you back up
an incompressible file on week one, back up the exact same file again on week two, and
never back it up again, this method provides a 2:1 data deduplication ratio. A more likely
scenario is that a portion of your data changes from backup to backup so that your data
deduplication ratio changes over time. With data deduplication, you can minimize your
storage requirements.
Data deduplication can provide greater data reduction and storage space savings than other
existing technologies.
Figure 6-13 shows the concept of data deduplication.
Data deduplication can reduce your storage requirements but the benefit you derive is
determined by your data and your backup policies. Workloads with a high database content
have the highest data deduplication ratios. However, product functions, such as IBM Tivoli
Storage Manager Progressive Incremental or Oracle Recovery Manager (RMAN), can reduce
the data deduplication ratio. Compressed, encrypted, or otherwise scrambled workloads
typically do not benefit from data deduplication. Good candidates for data deduplication are
text files, log files, uncompressed and non-encrypted database files, email files (PST, DBX,
and IBM Domino), and Snapshots (Filer Snaps, BCVs, and VMware images).
147
148
149
Cloud storage provides a ready-made data storage solution that helps in the following areas:
150
Chapter 7.
151
152
153
Fibre Channel
Internet Small Computer System Interface (iSCSI)
Serial Storage Architecture (SSA)
Fibre Channel over IP (FCIP)
We do not see many of these devices today, and they are considered historical devices.
154
In FC-AL, all devices on the loop share the bandwidth. The total number of devices that might
participate in the loop is 126, without using any hubs or fabric. For practical reasons, however,
the number tends to be limited to no more than 10 - 15.
Hubs are typically used in a SAN to attach devices or servers that do not support switched
fabric-only FC-AL. They might be unmanaged hubs, managed hubs, or switched hubs.
Unmanaged hubs serve as cable concentrators and as a means to configure the arbitrated
loop that is based on the connections that it detects. When one of the interfaces, typically a
gigabit interface converter (GBIC), on the hub senses that no cable is connected, that
interface shuts down. The hub port is then bypassed as part of the arbitrated loop
configuration.
Managed hubs offer all of the benefits of unmanaged hubs, but in addition, they offer the
ability to manage them remotely by using Simple Network Management Protocol (SNMP).
By using FC-AL, you can connect many servers and storage devices without using costly
Fibre Channel switches. FC-AL is not used much today because switched fabrics now lead in
the Fibre Channel market.
Switched hubs
Switched hubs allow devices to be connected in their own arbitrated loop. These loops are
then internally connected by a switched fabric.
155
A switched hub is useful to connect several FC-AL devices together, but to allow them to
communicate at full Fibre Channel bandwidth rather than all share the bandwidth.
Switched hubs are typically managed hubs.
FC-AL: Originally, FC-AL was described as SCSI on steroids. Although FC-AL has the
bandwidth advantage over SCSI, it does not come anywhere close to the speeds that can
be achieved and sustained on an individual port basis in a switched fabric. For this reason,
FC-AL implementations are, by certain observers, considered historical SANs.
Name service
Fabric control
Time service
Automatic discovery and registration of host and storage devices
Rerouting of frames, if possible, in a port problem
Storage services (virtualization, replication, and extended distances)
156
7.2.6 Multiplexers
Multiplexing is the process of simultaneously transmitting multiple signals over the same
physical connection. Common types of multiplexing are used for fiber optic connections that
are based on either time or wavelength:
Time-division multiplexing (TDM)
Wavelength division multiplexing (WDM)
Dense wavelength division multiplexing (DWDM)
When you use multiplexers in a SAN environment, more parameters in the SAN switch
configuration might be needed to ensure correct load balancing. Therefore, check with your
SAN switch vendor for preferred practices.
Multiplexers: Usually multiplexers are transparent to the SAN fabric. If you are
troubleshooting an ISL that covers a long distance, remember that the multiplexer, if
installed, plays an important role in that path.
7.3 Components
Many components must interoperate to create a SAN. We identify several common
components.
157
Frame filtering
Frame filtering is a feature that enables devices to provide zoning functions with finer
granularity. Frame filtering can be used to set up port-level zoning, worldwide name (WWN)
zoning, device-level zoning, protocol-level zoning, and logical unit number (LUN)-level zoning.
Frame filtering is commonly carried out by an ASIC. After you set up the filter, the complicated
function of zoning and filtering can be achieved at wire speed.
7.3.3 SerDes
The communication over a fiber, whether optical or copper, is serial. Computer busses,
however, use parallel busses. Therefore, Fibre Channel devices must be able to convert
between the two types. For this conversion, the devices use a serializer/deserializer, which is
commonly referred as a SerDes.
158
Layer 1 of the open systems interconnection (OSI) model is the layer at which the physical
transmission of data occurs. The unit of transmission at Layer 1 is a bit. This section explains
the common concepts of the Layer 1 level.
Single-mode fiber
Single-mode fiber (SMF) uses long wavelength light to transmit data and requires a cable with
a small core for transmission (Figure 7-3 on page 159). The core diameter for single-mode
159
cabling is
nine microns in diameter (Figure 7-4).
Multimode fiber
Multimode fiber (MMF) uses short wavelength light to transmit data and requires a cable with
a larger core for transmission (see Figure 7-3 on page 159). The core diameter for multimode
cabling can be 50 or 62.5 microns in diameter (Figure 7-5).
Outer Coating
250 micron diameter
Cladding
125 micron diameter
Core
50-micron or 62.5-micron diameter
The color of the outer coating is sometimes used to identify whether a cable is a multimode or
single-mode fiber cable, but the color is not a reliable method. The Telecommunications
Industry Association-598C (TIA-598C) standard suggests a yellow outer coating for single
mode fiber and an orange outer coating for multimode fiber for civilian applications. This
guideline is not always implemented, as illustrated in Figure 7-6, which shows a blue cable.
The reliable method is to look at the specifications of the cable that are printed on the outer
coating of the cabling. See also Figure 7-7 on page 161 and Figure 7-8 on page 161.
160
Copper cabling
When we refer to copper cabling, we mean that the material that is used to transfer the
signals is made of copper. The most common copper wire is the twisted-pair cable that is
used for normal Ethernet. This type of cabling is explained in more depth in the following
section.
Twisted-pair cabling
Twisted-pair copper cabling is a common media for Ethernet networking installations.
Twisted-pair cabling is available as unshielded twisted pair (UTP) or shielded twisted pair
(STP). This shielding helps prevent electromagnetic interference.
Several categories of twisted-pair cabling are available (Table 7-1). These categories indicate
the signaling capabilities of the cabling.
Table 7-1 TIA/Electronic Industries Alliance (EIA) cabling categories
TIA/EIA cabling category
Cat 1
Telephone or ISDN
Cat 2
4 Mb Token Ring
Cat 3
10 Mb Ethernet
Cat 4
16 Mb Token Ring
Cat 5
100 Mb Ethernet
Cat 5e
1 Gb Ethernet
Cat 6
10 Gb Ethernet
Short Distance - 55 m (180 ft)
Cat 6a
10 Gb Ethernet
161
The connector that is used for Ethernet twisted-pair cabling is likely the connector that most
people recognize and associate with networking, which is the RJ45 connector. Figure 7-9
shows this connector.
Twisted-pair cabling contains four pairs of wire inside the cable (Figure 7-10).
An Ethernet that is operating in 10/100 Mb mode uses only two pairs: pairs 1-2 and 3-6. An
Ethernet that is operating in 1 Gb mode uses all four pairs: pairs 1-2, 3-6, 4-5, and 7-8.
Distances up to 100 meters (328.08 feet) are supported.
Damaged twisted pair: If a twisted-pair cable is damaged so that pair 4-5 or pair 7-8 is
unable to communicate, the link is unable to communicate in 1 Gbps mode. If the devices
are set to auto-negotiate speed, the devices successfully operate in 100 Mbps mode.
Supported maximum distances of cabling segment: The actual maximum distances of
a cabling segment that are supported vary due to multiple factors, such as vendor support,
cabling type, electromagnetic interference, and the number of physical connections in the
segment.
162
Twinax cabling
Twinax cables were used by IBM for many years, but they were recently reintroduced to the
market as a transport media for 10 Gb Ethernet. One of the greatest benefits of a twinax
cable is its low power consumption. Also, this cable costs less than standard fiber cables. The
downside is the limited capability to connect over long distance.
Connector types
The most common connector type for fiber optic media that is used in networking today is the
LC connector (Figure 7-11).
Other types of connectors are the SC connector (Figure 7-12) and the ST connector (not
shown).
7.4.2 Transceivers
A transceiver or transmitter/receiver is the fiber optic port of a device where the fiber optic
cables connect. Occasionally, a device might include an integrated transceiver, which limits
the flexibility in the type of cabling that you can use. Most devices provide a slot to insert a
modular transceiver, providing flexibility so that you can select either single or multimode
implementations.
163
Certain equipment might use a larger transceiver that is known as a Gigabit Interface
Converter (GBIC) (Figure 7-13). As technology advances, smaller transceivers are
introduced. These smaller transceivers provide much higher port density, such as small
form-factor pluggables (SFPs), 10 Gigabit SFP+, 10 Gigabit SFP-XFP, and Quad SFP
(QSFP).
Figure 7-14 From left to right: SFP-MMF, SFP-SMF, SFP+-MMF, XFP-MMF, and XFP-SMF
164
165
Several manufacturers offer HBAs. The choice of the HBA is an important consideration when
you plan a SAN. HBAs might include more than one port. They can be supported by certain
equipment and not other equipment. HBAs might include parameters to tune the system.
Many other features are available. HBAs also have a certain number of buffer-to-buffer
credits.
Important: If you are considering the use of an HBA with multiple virtual machines behind
it, the choice of an HBA is a critical decision.
167
7.5.1 Cascading
Expanding the fabric is called switch cascading, or cascading. Cascading is interconnecting
Fibre Channel switches and directors by using ISLs. By cascading switches, the following
benefits are possible for a SAN environment:
The fabric can be seamlessly extended. Additional switches can be added to the fabric
without powering down the existing fabric.
You can increase the distance between various SAN participants easily.
By adding switches to the fabric, you increase connectivity by providing more available
ports.
Cascading provides high resilience in the fabric.
With ISLs, you can increase the bandwidth. The frames between the switches are
delivered over all available data paths. Create more ISLs to increase the speed of the
frame delivery. However, be careful to ensure that you do not introduce a bottleneck.
When the fabric grows, the name server is fully distributed across all of the switches in the
fabric.
With cascading, you also provide greater fault tolerance within the fabric.
7.5.2 Hops
When Fibre Channel traffic traverses an ISL, this process is known as a hop. Or, to state it
another way, traffic that goes from one E_port over an ISL to another E_port is one hop. ISLs
are created by connecting an E_port to an E_port. Figure 7-18 on page 169 shows an
illustration of the hop count from server to storage.
The hop count is limited. This limit is set by the fabric operating system. The limit is used to
derive a frame hold time value for each switch. This value is the maximum amount of time that
a frame can be held in a switch before the frame is dropped, or the fabric indicates that it is
too busy. The hop count limits must be investigated and considered in any SAN design work
because the hop count limit significantly affects the proposal.
168
For example, as shown in Figure 7-18, if a server needs to connect to its storage through
multiple switches, FSPF routes all traffic from this server to its storage through switch A
directly to switch C. This path is taken because it has a lower cost than traveling through more
hops through switch B.
Figure 7-18 shows hops in a fabric.
169
Blocking
Blocking means that the data does not get to the destination. Blocking is not the same as
congestion because with congestion, data is still delivered, but delayed. Currently, almost all
Fibre Channel switches are created by using non-blocking architecture.
Non-blocking
A non-blocking architecture is used by most switch vendors. Non-blocking switches enable
multiple connections that travel through the switch at the same time. Figure 7-19 shows this
concept.
Switch B
Switch A
A
Non-blocking
Blocking
170
7.5.5 Latency
Typically, in the SAN world, latency is the time that it takes for a Fibre Channel frame to
traverse the fabric. When we describe the SAN, the latency in a SAN is rarely considered
because it is in the low microsecond range. This concept is sometimes confused with disk
latency, which is the measure of how quickly or slowly a storage target completes a read or
write request that is sent from the server. However, when we describe long distances, all
latency, both storage and SAN, plays a significant role.
Latency increases as the number of ISLs increases because the Fibre Channel frame must
traverse the fabric by using ISLs. By fabric, we mean the Fibre Channel components and any
latency discussion that relates to the SAN. Usually, the time that is taken is expressed in
microseconds, which indicates the performance characteristics of the SAN fabric. Latency is
often provided at a switch level, and sometimes at a fabric level.
7.5.6 Oversubscription
Another aspect of data flow is the fan-in ratio, which is also called the oversubscription ratio
and frequently the fan-out ratio from the storage devices perspective, both in terms of host
ports to target ports, and the device to the ISL. This ratio is the number of device ports that
need to share a single port.
For example, two servers, each of which is equipped with a 4 Gb port (4+4=8 Gb) are both
communicating with a storage device through a single 4 Gb port, which provides a 2:1 ratio.
Therefore, the total theoretical input is higher than the input that the port can provide.
Figure 7-20 on page 172 shows a typical oversubscription through an ISL.
Oversubscription can occur on storage device ports and ISLs. When you design a SAN, it is
important to consider the possible traffic patterns to determine the possibility of
oversubscription. An oversubscription might result in degraded performance. You can
overcome the oversubscription of an ISL by adding an ISL between the switches to increase
the bandwidth. Oversubscription to a storage device might be overcome by adding more ports
from the storage device to the fabric.
Oversubscription: Vendors differ in how they practice utilization on their stated overall
bandwidth for each chassis. Vendors use both storage port and ISL oversubscription.
Verify the oversubscription preferred practices with your switch vendor.
7.5.7 Congestion
Oversubscription leads to a condition that is called congestion. When a node is unable to use
as much bandwidth as it wants, because of contention with another node, congestion occurs.
A port, link, or fabric can be congested. This condition normally affects the application directly
and results in poor performance.
Congestion can be difficult to detect because it can also relate directly to buffer-to-buffer
credit starvation in the switch port. Therefore, when you look at the data throughput from the
switch, you see normal or less traffic flowing through the ports. However, the server I/O is
unable to perform because the data cannot be transported because of a lack of
buffer-to-buffer credits.
171
0 .1 G b p s
0 .1 G b p s
B
1
1 .7 5 G b p s 1 .2 5 G b p s
E
2
F
2
A
0 .1 G b p s
4x2=8 G bps
In Figure 7-20, six computers access three storage devices. Computers A, B, C, and D
communicate with storage G. Server E communicates with storage H. Server F uses disks in
storage device I.
The speeds of the links are shown in Gbps, and the target throughput for each computer is
shown. If we allow FSPF alone to decide the routing, servers D and E might both use the
same ISL. This situation leads to oversubscription and therefore congestion because
1.7+1.75 is greater than 2.
If all of the ISLs are gathered into a trunk, effectively they are a single, large ISL. They appear
as an 8 Gbps ISL. This bandwidth is greater than the total requirement of all of the servers. In
fact, the nodes require an aggregate bandwidth of 5 Gbps. Therefore, one of the ISLs might
fail and you still have sufficient bandwidth to satisfy the needs.
When the nodes come up, FSPF simply sees one route, and the nodes are all assigned a
route over the same trunk. The fabric operating systems in the switches share the load over
the actual ISLs, which combine to make up the trunk. This process is performed by
distributing frames over the physical links and then reassembling them at the destination
switch so that an in-order delivery can be assured, if necessary. And, to FSPF, a trunk is
displayed as a single, low-cost ISL.
172
Chapter 8.
Management
Management is a key issue behind the concept of infrastructure simplification. The ability to
manage heterogeneous systems at different levels as though they were a fully integrated
infrastructure is a goal that many vendors and developers strive to achieve. Another goal is to
offer the system administrator a unified view of the whole storage area network (SAN).
In this chapter, we look at several initiatives in the field of SAN management. These solutions
incrementally smooth the way toward infrastructure simplification.
173
Capacity management
Device management
Fabric management
Proactive monitoring
Fault isolation and troubleshooting
Centralized management
Remote management
Performance management
Security and standard compliance
174
Figure 8-1 shows in-band and out-of-band models. These models are not mutually exclusive.
In many environments, you might want a combination of the models.
Server
Server
Server
Server
Server
Server
Data
Storage
Storage
In-band
Out-of-band
Data + Control
The in-band approach is simple to implement. This approach requires no dedicated channels
(other than LAN connections). It offers inherent advantages, such as the ability for a switch to
initiate a SAN topology map with queries to other fabric components. However, if the Fibre
Channel transport fails, the management information cannot be transmitted. Therefore, the
access to devices and the ability to detect, isolate, and recover from network problems are
lost. This problem can be minimized by a provision of redundant paths between devices in the
fabric.
In-band management allows attribute inquiries on storage devices and configuration changes
for all elements of the SAN. Because in-band management is performed over the SAN,
administrators are not required to manage any additional connections.
Conversely, out-of-band management does not rely on the storage network; its major
advantage is that management commands and messages can be sent even if a loop or fabric
link fails. Integrated SAN management facilities are more easily implemented. However,
unlike in-band management, it cannot automatically provide SAN topology mapping.
In summary, in-band management offers these major advantages:
Chapter 8. Management
175
ISL
Switch
Switch
UNIX
Red Fabric
Storage
Windows
Storage
Management
Console
ISL
VMware
Tape
Switch
Switch
Green Fabric
The SAN storage level consists of the storage devices that integrate the SAN, such as disks,
disk arrays, tapes, and tape libraries. Because the configuration of a storage resource must
be integrated with the configuration of the servers logical view of the storage resources, the
SAN storage level management can also span both storage resources and servers.
176
Connections
Naming conventions
Device serial numbers
Worldwide names (WWNs)
Zoning
System applications
A responsible person must be in charge of maintaining this infrastructure and monitoring the
SANs health status.
Chapter 8. Management
177
178
IBM developed its storage management solutions based on the CIMOM architecture
(Figure 8-3).
Chapter 8. Management
179
This protocol is widely supported by LAN/wide area network (WAN) routers, gateways, hubs,
and switches. SNMP is the predominant protocol that is used for multivendor networks.
Device status information (vendor, machine serial number, port type and status, traffic, errors,
and so on) can be provided to an enterprise SNMP manager. A device can generate an alert
by SNMP in an error condition. The device symbol, or icon, which is displayed on the SNMP
manager console, can be changed to red or yellow, or any warning color, and messages can
be sent to the network operator.
Out-of-band developments
SNMP MIBs are implemented for SAN fabric elements that allow out-of-band monitoring. The
ANSI Fibre Channel Fabric Element MIB provides significant operational and configuration
information about individual devices. The emerging Fibre Channel Management MIB provides
more link table and switch zoning information that can be used to derive information about the
physical and logical connections between individual devices.
Application programming interfaces (APIs) are one way to help this simplification become a
reality. Many vendors make the API of their product available for other vendors to make it
possible for common management in the SAN. This openness allows the development of
upper-level management applications that interact with multiple vendor devices and offer the
system administrator a single view of the SAN infrastructure.
180
Chapter 8. Management
181
182
Cost-effectiveness
An open approach
Device management
Fabric management
Proactive monitoring
Fault isolation and troubleshooting
Centralized management
Remote management
Adherence to standards
Resource management
Secure access
Standards compliance
8.3.1 Operations
When we describe management, it automatically includes the operational aspects of the
environment. The SAN administrators are responsible for all configuration of the SAN
switches.
Typically, the initial design and creation of a SAN environment includes only a handful of
servers and few storage systems. However, the environment grows and new technology
needs to be added. At this stage, it tends to get more complicated. Therefore, it is necessary
to ensure that comprehensive documentation exists that describes all aspects of the
environment. And, the documentation needs to be reviewed regularly to ensure that it is
current.
The following standards and guidelines need to be documented:
Zoning standards:
How to create zones by using preferred practices
Naming standards that are used in the SAN configuration
Aliases used
Volume/LUN allocation standards:
Volume characteristics and their uses
Allocation rules
Incident and problem guidelines: How to react in case of an incident.
Roles and responsibilities: Roles and responsibilities within the team.
SAN and storage installation preferred practices: Agreed-to process to install and
configure the equipment.
SAN and storage software and firmware upgrade roadmaps:
High-level overview of how to ensure that the environment is kept current
Change schedules
Monitoring and performance guidelines, such as defining the components, software, and
processes that are monitored and explaining how to handle exceptions.
Chapter 8. Management
183
8.4.1 b-type
The b-type family switch management framework is designed to support the widest range of
solutions, from the small workgroup SANs up to large enterprise SANs. The software that
Brocade (the IBM original equipment manufacturer (OEM) partner) provides is called Data
Center Fabric Manager (DCFM) and Brocade Network Advisor (BNA). This software was
added to the IBM portfolio as IBM Data Center Fabric Manager and IBM Network Advisor.
The portfolio also includes Fabric Vision.
You can use the following tools with b-type SANs to centralize control and enable the
automation of repetitive administrative tasks:
Web Tools
A built-in web-based application that provides administration and management functions
on an individual switch basis.
Data Center Fabric Manager (DCFM)
A client/server-based external application that centralizes the management of
IBM/Brocade multiprotocol fabrics within and across data centers, including support for
Fibre Channel over Ethernet (FCoE) and Converged Enhanced Ethernet (CEE).
Fabric Watch
A Fabric OS built-in tool that allows the monitoring of key switch elements: power supplies,
fans, temperature, error counters, and so on.
SNMP
A feature that enables storage administrators to manage storage network performance,
locate and solve storage network problems, and plan for storage network growth.
You can use the following management interfaces to monitor fabric topology, port status,
physical status, and other information to help with system debugging and performance
analysis:
You can use all of these management methods either in-band (Fibre Channel) or out-of-band
(Ethernet), except for SES, which can be used for in-band only.
Additionally, Fabric Vision technology is an advanced hardware and software solution that
combines capabilities from our b-type Gen 5 Fibre Channel solutions, Fabric OS (FOS), and
IBM Network Advisor to help administrators avoid problems, maximize application
performance, and reduce operational costs through innovative diagnostic, monitoring, and
management technology.
Fabric Vision technology includes ClearLink Diagnostics, Monitoring and Alerting Policy Suite
(MAPS), Flow Vision, Bottleneck Detection, At-a-glance dashboard, and Forward Error
Correction (FEC).
For more information about IBM Network Advisor, see this website:
http://www.ibm.com/systems/storage/san/b-type/na/
For more information about Fabric Vision, see this website:
https://ibm.biz/BdH2nB
184
8.4.2 Cisco
Cisco Data Center Network Manager (DCNM) is a management system for the Cisco Unified
Fabric. With Cisco DCNM, you can provision, monitor, and troubleshoot the data center
network infrastructure. Cisco DCNM provides visibility and control of the unified data center
so that you can optimize for the quality of service (QoS) that is required to meet service level
agreements (SLAs).
Cisco DCNM increases overall data center infrastructure uptime and reliability, improving
business continuity. It provides a robust framework and comprehensive feature set that meets
the routing, switching, and storage administration needs of data centers. Cisco DCNM
streamlines the provisioning for the unified fabric and monitors the SAN and LAN
components. Cisco DCNM provides a high level of visibility and control through a single
web-based management console for Cisco Nexus, Cisco MDS, and Cisco Unified Computing
System products.
Cisco DCNM also supports the installation of the Cisco DCNM for SAN and Cisco DCNM for
LAN components with a single installer.
Cisco DCNM offers these capabilities:
Configures and manages the fabric on multiple efficient levels
Groups multiple SAN objects and SAN management functions intelligently to provide ease
and time efficiency in administering tasks
Identifies, isolates, and manages SAN events across multiple switches and fabrics
Provides drill-down capability to individual SAN components through tightly coupled
Web Tools and Fabric Watch integration
Discovers all SAN components and views so that you can see the real-time state of all
fabrics
Provides the multi-fabric administration of secure Fabric OS SANs through a single
encrypted console
Monitors ISLs
Manages switch licenses
Performs fabric stamping
For more information about Cisco DCNM, see this website:
https://ibm.biz/BdH2G5
Chapter 8. Management
185
Traditionally, multipathing software was supplied by each vendor to support that vendors
storage arrays. The multipathing software often is embedded in the operating system. This
approach offers a server-centric approach to multipathing that is independent of the storage
array. This approach is often easier to implement from a testing and migration viewpoint.
Difference between storage and a SAN: It is important to understand the key difference
between a SAN and storage, although sometimes they are referred to as one.
ISL
Switch
Switch
UNIX
Red Fabric
Windows
Storage
ISL
Switch
VMware
Switch
Green Fabric
The IBM multipathing software is IBM Subsystem Device Driver (SDD), which offers the
following benefits:
186
When you determine the number of paths to configure to each volume, never exceed the level
that is supported by the storage device. When you implement zoning to a storage device, you
must decide the number of paths.
For detailed information about multipath drivers for IBM storage, see the following website:
http://www.ibm.com/support/docview.wss?rs=540&context=ST52G7&q=ssg1*&uid=ssg1S7000
303&loc=en_US&cs
We provide a multipathing example in separate scenarios. In Figure 8-5 on page 186, the
servers are connected to the SAN with two HBAs. The HBAs access the volumes through two
storage ports on the storage device. This access is controlled by zoning, which provides four
working paths for each server to their volumes: two from the Red Fabric and two from the
Green Fabric.
Figure 8-6 shows a single path failure, which is indicated by the STOP sign.
ISL
UNIX
Switch
Switch
Red Fabric
Windows
Storage
ISL
VMware
Switch
Switch
Green Fabric
In Figure 8-6, the Windows server lost connectivity to the SAN and cannot access the Red
Fabric. However, the Windows server has working paths through the Green Fabric. All other
servers are running without any issues.
Chapter 8. Management
187
ISL
UNIX
Switch
Switch
Red Fabric
Windows
Storage
ISL
VMware
Switch
Switch
Green Fabric
In Figure 8-7, our servers cannot access a switch. Working paths still exist from the servers
through the Green Fabric.
188
Figure 8-8 shows that a link from the storage device to a switch failed in the Red Fabric.
ISL
UNIX
Switch
Switch
Red Fabric
Windows
Storage
ISL
VMware
Switch
Switch
Green Fabric
In Figure 8-8, the storage device lost one of four connections. One connection to the Red
Fabric does not function. Therefore, all servers that use the same storage port now see three
working paths out of four possible paths. All servers that are zoned to the failed storage port
are affected.
Chapter 8. Management
189
Figure 8-9 shows the storage device lose access to the Red Fabric.
ISL
UNIX
Switch
Switch
Red Fabric
Windows
Storage
VMware
ISL
Switch
Switch
Green Fabric
Figure 8-9 Storage device that lost two out of the four connections to the SAN
In Figure 8-9, our storage device lost access to the Red Fabric. All devices in the Red Fabric
are running normally. Only two specific storage ports failed. Our servers have only two
working paths through the Green Fabric. This configuration affects all servers that are zoned
to these storage ports.
190
ISL
UNIX
Switch
Switch
Red Fabric
Windows
Storage
ISL
VMware
Switch
Switch
Green Fabric
In Figure 8-10, we lost our storage device. No paths to any volumes on this device are
available. No data is accessible. All servers that are zoned to this storage device are severely
affected and cannot access this storage device.
If we install a supported version of a multipath driver correctly on all of the servers, we survive
all scenarios except the last scenario, in which only a minimum impact occurs.
Chapter 8. Management
191
192
Chapter 9.
Security
In this chapter, we provide an overview of the need for security. We describe the available
techniques and several key points.
193
194
The secure system must challenge the user (typically with a password)
so that this user is identified.
Authorization
After the system identifies a user, the system knows what this user is
allowed to access and what they are not allowed to access.
The system must guarantee that the data is stored or processed within
its boundaries and that it is not altered or tampered with in any way.
The data security and integrity requirement aims to guarantee that data from one application
or system does not become overlaid, corrupted, or otherwise destroyed. This requirement
applies whether data is intentionally destroyed or destroyed by accident, either by other
applications or systems. This requirement might involve a form of authorization, and the
ability to fence off the data in one system from another system.
Chapter 9. Security
195
This data security necessity must be balanced with the requirement for the expansion of
SANs to enterprise-wide environments, with an emphasis on multiple platform connectivity.
True cross-platform data sharing solutions, as opposed to data partitioning solutions, are also
a requirement. Security and access control also must be improved to guarantee data integrity.
We overview several common data security approaches for the SAN environment. This list is
not meant to be an in-depth description. It is merely an attempt to acquaint you with the
technology and terminology that you are likely to encounter in a discussion about SAN
security.
Zoning
Initially, SANs did not have any zoning. It was an any-to-any communication. No real access
control mechanism protected storage that was used by one host from being accessed by
another host. When SANs grew, this drawback became a security risk as SANs became more
complex and ran more vital parts of the business. To mitigate the risk of unwanted cross
communication, zoning was invented to isolate communication to devices within the same
zone.
Persistent binding
Server-level access control is called persistent binding. Persistent binding uses configuration
information that is stored on the server. Persistent binding is implemented through the HBA
driver of the server. This process binds a server device name to a specific Fibre Channel
storage volume or logical unit number (LUN), through a specific HBA and storage port WWN.
Or, put in more technical terms, it is a host-centric way to direct an operating system to assign
certain Small Computer System Interface (SCSI) target IDs and LUNs.
196
Port binding
To provide a higher level of security, you can also use port binding to bind a particular device
(as represented by a WWN) to a specific port that does not allow any other device to plug into
the port.
Chapter 9. Security
197
198
Asymmetric
A symmetric cryptosystem follows a fairly straightforward philosophy: Two parties can securely
communicate if both parties use the same cryptographic algorithm and possess the same
secret key to encrypt and decrypt messages. This algorithm is the simplest and most efficient
way of implementing secure communication, if the participating parties are able to securely
exchange secret keys (or passwords).
Figure 9-1 shows symmetric encryption.
Chapter 9. Security
199
9.4.2 Data-in-flight
Also known as data-in-motion, this term generically refers to protecting information any time
that the data leaves its primary location, for example, when data is transmitted from the
source across any type of network to a target. To secure this transmission, we use
technologies, such as Secure Sockets Layer (SSL), Virtual Private Network (VPN), and
IP Security (IPSec) to assure data confidentiality. Then, we use other technologies, such as
digital certificates, message authentication codes, and keyed hashes, to ensure data integrity.
Data-in-flight is also information (data) that leaves the data center through, for example, an
open network or leased dark fiber.
All of these areas can be addressed with encryption-based technologies.
200
9.4.3 Data-at-rest
Protecting data as it resides on the storage media, disk, or tape is typically referred to as
protecting data-at-rest.
If encryption is used as part of the strategy for the protection of data-at-rest, this protection
also indirectly addresses the issue of displayed tape media. This issue is addressed because,
even if tapes fall into the wrong hands, the data that is stored on them is unreadable without
the correct key. These security measures assume that you enacted the appropriate key
management techniques.
To gain the needed security level, you build layers of security on your SAN. You first increase
the level of difficulty for an unauthorized user to even gain access to the data. You then
compound that with the fact that private data is not stored in human-readable form.
For more information about IBM System Storage Data Encryption, see IBM System Storage
Data Encryption, SG24-7797. For an example of how IBM implements encryption on the
IBM System Storage SAN Volume Controller, see Implementing the Storwize V7000 and the
IBM System Storage SAN32B-E4 Encryption Switch, SG24-7977.
If we look at the security aspect on its own, were focused on establishing a perimeter of
defense around system assets. Although securing access to our environments continues to
be an important part of security, the typical business cannot afford to lock down its entire
enterprise.
Chapter 9. Security
201
Open networks are now commonly used to connect clients, partners, employees, suppliers,
and their data. Although open networks offer significant advantages, they raise concerns
about how a business protects its information assets and complies with industry and
legislative requirements for data privacy and accountability. By using data encryption as a
part of the solution, many of these concerns can be mitigated.
202
Master key: Even with a key server, generally at least one encryption key, which is
normally called the master key (MK), must be maintained manually. For example, this
master key manages access to all other encryption keys. This master key encrypts the
data that is used by the key server to exchange keys.
Fundamentally, IBM Security Key Lifecycle Manager works by allowing administrators to
connect to storage devices and then create and manage keystores. These stores are secure
repositories of keys and certificate information that are used to encrypt and decrypt data, or to
use existing keystores already in place.
Over the course of the key lifecycle, all management functions, including creation,
importation, distribution, backup, and archiving, are easily accomplished. These functions can
be performed by using the lifecycle managers graphic interface, which can be accessed by
using any standard browser in the network.
IBM Security Key Lifecycle Manager therefore serves as a central point of control, unifying
key management even when different classes of storage devices are involved. For more
information about IBM Security Key Lifecycle Manager, see this website:
http://www.ibm.com/software/products/en/key-lifecycle-manager
Two security standards are important to ensuring the integrity of encryption products: FIPS
140 and Common Criteria. The official title for the standard Federal Information Processing
Standard 140 (FIPS-140) is Security Requirements for Cryptographic Modules. FIPS 140-2
stands for the second revision of the standard and was released in 2001. Common Criteria
has seven Evaluation Assurance Levels (EALs), which were defined in 1999. Together, these
standards support a small industry for certifying security products and ensuring the integrity
of encryption systems.
In-flight encryption
The in-flight encryption and compression feature of Fabric OS allows frames to be encrypted
or compressed at the egress point of an inter-switch link (ISL) between two b-type switches,
and then to be decrypted or extracted at the ingress point of the ISL. This feature uses
port-based encryption and compression. It is supported on 16 Gbps E_ports only.
Note: The ports can run at 2 Gbps, 4 Gbps, 8 Gbps, 10 Gbps, or 16 Gbps.
The purpose of encryption is to provide security for frames while they are in flight between
two switches. The purpose of compression is for better bandwidth use on the ISLs, especially
over long distances. An average compression ratio of 2:1 is provided. Frames are never left in
an encrypted or compressed state when they are delivered to an end device, and both ends
of the ISL must terminate at 16 Gbps ports.
For more information, see the Metro Cloud Connectivity: Integrated Metro SAN Connectivity
in Gen 5 Fibre Channel Switches white paper:
http://www.brocade.com/content/dam/common/documents/content-types/whitepaper/broca
de-metro-cloud-connect-wp.pdf
Chapter 9. Security
203
Encryption at rest
The b-type fabric-based encryption solutions work transparently with heterogeneous servers,
tape libraries, and storage subsystems. Although host-based encryption works only for a
specified operating system and storage-based encryption works only for a specific vendor,
b-type products are deployed in the core of the fabric to encrypt Fibre Channel-based traffic.
Users deploy b-type encryption solutions through either the FS8-18 Encryption Blade or the
2U, rack-mounted IBM SAN32B-E4 Encryption Switch.
The Device Encryption Key (DEK) is important. Because the DEK is needed to encrypt and
decrypt the data, it must be random and 256 bits. B-type encryption devices use a True
Random Number Generator (TRNG) to generate each DEK. For encrypting data that is
destined for a disk drive, one DEK is associated with one logical unit number (LUN).
The Institute of Electrical and Electronic Engineers 1619 (IEEE 1619) standard on encryption
algorithms for disk drives is known as AES256-XTS. The encrypted data from the
AES256-XTS algorithm is the same length as the unencrypted data. Therefore, the b-type
encryption device can encrypt the data, block by block, without expanding the size of the data.
The key management is performed by using external software, such as IBM Security Key
Lifecycle Manager.
204
In-flight encryption
Cisco TrustSec Fibre Channel Link Encryption is an extension of the FC-SP standard. It uses
the existing FC-SP architecture. Fibre Channel data that travels between E_ports on 8 Gbps
modules is encrypted. Cisco uses the 128-bit Advanced Encryption Standard (AES)
encryption algorithm and enables either AES-Galois/Counter Mode (AES-GCM) or
AES-Galois Message Authentication Code (AES-GMAC). AES-GCM encrypts and
authenticates frames, and AES-GMAC authenticates only the frames that are passed
between the two peers.
Encryption is performed at line rate by encapsulating frames at egress with encryption by
using the GCM authentication mode with 128-bit AES encryption. At ingress, frames are
decrypted and authenticated with integrity checks.
Two primary use cases for Cisco TrustSec Fibre Channel Link Encryption exist. In the first use
case, clients are communicating outside the data center over native Fibre Channel (for
example, dark fiber, Coarse Wavelength-Division Multiplexing (CWDM) or Dense
Wavelength-Division Multiplexing (DWDM). In the second use case, encryption is performed
within the data center for security-focused clients, such as defense and intelligence services.
Chapter 9. Security
205
Encryption at rest
Cisco uses Storage Media Encryption (SME), which protects data at rest on heterogeneous
tape drives, virtual tape libraries (VTLs), and disk arrays, in a SAN environment by using
highly secure IEEE Advanced Encryption Standard (AES) algorithms.
Encryption is performed as a transparent Fibre Channel fabric service, which greatly
simplifies the deployment and management of sensitive data on SAN-attached storage
devices. Storage in any virtual SAN (VSAN) can make full use of Cisco SME.
Secure lifecycle key management is included with essential features, such as key archival,
shredding, automatic key replication across data centers, high-availability deployments, and
export and import for single-site and multiple-site environments. Provisioning and key
management for Cisco SME are both integrated into Cisco Fabric Manager and Data Center
Network Manager (DCNM). No additional software is required for key management.
Figure 9-6 shows the SME architecture.
206
Chapter 9. Security
207
208
10
Chapter 10.
Solutions
The added value of a storage area network (SAN) lies in the use of its technology to provide
tangible and desirable benefits to the business. These benefits are provided by the use of
fast, secure, reliable, and highly available networking solutions. Benefits range from increased
availability and flexibility to more functionality that can reduce application downtime.
In this chapter, we provide a description of general SAN applications, and the types of
components that are required to implement them. Far more complexity exists than is
presented here. For instance, this text does not cover how to choose one switch over another,
or how many inter-switch links (ISLs) are necessary for a specific SAN design. These
strategic decisions must be always considered by experienced IT architects, and these
decisions are beyond the intended scope of this book. We introduce the basic principles and
key considerations to choose an optimal solution for SAN deployments.
209
10.1 Introduction
With the continued development of communication and computing technologies and
products, SANs are becoming much more complex. We are not referring to merely a simple
fiber-optic connection between SAN devices. Examples of these devices include SAN
switches, routers, tape drives, disk device subsystems, and target host systems that use
standard Fibre Channel host bus adapters (HBAs). Technology moved beyond those
solutions and continues to do so.
Today, businesses are looking for solutions that enable them to increase the data transfer rate
within the most complex data centers. Businesses also want solutions that provide high
availability of managed applications and systems, implement data security, and provide
storage efficiency. At the same time, businesses want to reduce the associated costs and
power consumption.
Organizations must find a smooth, effective, and cost-efficient way to upgrade their current or
traditional SAN infrastructure. The upgraded infrastructure provides a less complex and more
powerful and flexible data center of the next generation.
SAN solutions can be classified into many categories. We chose to classify our SAN solutions
as infrastructure simplification, business continuity, and information lifecycle management
(ILM). In the following topics, we describe the use of basic SAN design patterns to build
solutions for different requirements. These requirements range from simple data movement
techniques that are frequently employed as a way to improve business continuity, up to
sophisticated storage-pooling techniques that are used to simplify complex infrastructures.
Before SAN solutions and requirements are described, we present basic principles to
consider when you plan a SAN implementation or upgrade.
10.2.1 Connectivity
Connecting servers to storage devices through a SAN fabric is often the first step in a phased
SAN implementation. Fibre Channel attachments offer the following benefits:
Extended connection distances (sometimes called remote storage)
Enhanced addressability
Improved performance by running Small Computer System Interface (SCSI) over
Fibre Channel
Many implementations of Fibre Channel technology are simple configurations that remove
many of the restrictions of the existing storage environments. With these implementations of
Fibre Channel technology, you can build one common physical infrastructure. The SAN uses
common cabling to the storage and other peripheral devices.
210
211
Data migration
One of the critical tasks for a SAN administrator is to move data between two independent
SAN infrastructures. The administrator might move data from an old storage system that is
being discontinued to the new enterprise and high-performance disk system. Two basic
scenarios exist. SANs are independent and cannot be interconnected even if they reside in
the same data center, and the disk systems can be cross-connected through SAN switches.
In Figure 10-1, the storage administrator is challenged to migrate data to the newly deployed,
high performance disk storage system without interrupting the clients critical SAP
applications. Luckily, we can manage both source and target storage systems. These
systems are configured to communicate through SAN switches. Disk copy services can
replicate specific LUNs from the old storage device to the new storage device without
affecting the performance of the SAP application.
212
In addition, you can use this procedure to prepare a standby application server that connects
to the replicated disk LUNs. Or, you can use this procedure to replace the old server
hardware where the SAP application is running, with the minimum outage that is necessary to
switch the application over to the prepared server.
The application server is connected to both SANs by using independent HBAs. Application
owners and SAN2 administrators analyze the current disk structure that is assigned from the
source storage system. The SAN2 administrator assigns the same disk capacity to the
application server. The application or system owner then migrates the data from the source
disk to the target disk. This migration is performed manually by using the operating system
functions. (The application is offline.) Or, disk mirroring must be enabled.
When the data is synchronized between the source and target disks, the mirror can be
broken, source disks can be unassigned, and the source storage system can be
disconnected from the application server. The disadvantage of this solution is a significant I/O
operation on the source and target LUNs that can potentially affect the performance of critical
applications.
213
214
One logical place to use 16 Gbps is in the inter-switch link (ISL) scenario. This scenario has
two advantages. The increased speed between switches is an obvious advantage. And, the
increased bandwidth might mean that you need fewer ISLs. If fewer ISLs are required, it might
be possible to reassign ISLs to attach hosts or storage.
Another consideration is cost. IT architects and investors must evaluate their current SAN
solutions in their data centers and make strategic decisions to determine whether it is
beneficial to continue with the upgrade to a dedicated Fibre Channel solution that is running
16 Gbps devices. Or, the architects and investors must determine whether now is the right
time to consider an upgrade to converged networks to use, for example, FCoE. Many
products are available on the market that support these transformations and transitions and
protect client investments for the future.
215
Disk pooling
Disk pooling allows multiple servers to use a common pool of SAN-attached disk storage
devices. Disk storage resources are pooled within a disk subsystem or across multiple IBM
and non-IBM disk subsystems. And, capacity is assigned to independent file systems that are
supported by the operating systems on the servers. The servers are potentially a
heterogeneous mix of UNIX, Microsoft Windows, and even IBM z Systems servers.
Storage can be dynamically added to the disk pool and assigned to any SAN-attached server
when and where necessary. This function provides efficient access to shared disk resources
without a level of indirection that is associated with a separate file server. This scenario is
possible because storage is effectively directly attached to all of the servers, and efficiencies
of scalability result from the consolidation of storage capacity.
When storage is added, you can use zoning to restrict access to the added capacity. Because
many devices (or LUNs) can be attached to a single port, access can be further restricted by
using LUN-masking. You can use LUN masking to specify who can access a specific device
or LUN.
216
You can attach and detach storage devices under the control of a common administrative
interface. Storage capacity can be added without stopping the server, and the storage
capacity can be available to the applications immediately.
Figure 10-3 shows an example of disk storage pooling across two servers.
AIX
Windows
SAN
Fabric
Storage A
Storage B
In Figure 10-3, one server is assigned a pool of disks that are formatted to the requirements
of the file system, and the second server is assigned another pool of disks, possibly in
another format. The third pool might be the space that is not yet allocated, or the third pool
can be a preformatted disk for future use. Again, all of the changes in the disk structure can
be performed dynamically without interruption to the service.
Tape pooling
Tape pooling addresses a problem in an open systems environment where multiple servers
are unable to share tape resources across multiple hosts. Older methods of sharing a device
between hosts consist of either manually switching the tape device from one host to the other,
or writing applications that communicate with connected servers through distributed
programming.
Tape pooling allows applications on one or more servers to share tape drives, libraries, and
cartridges in a SAN environment in an automated, secure manner. With a SAN infrastructure,
each host can directly address the tape device as though the tape device is connected to all
of the hosts.
Tape drives, libraries, and cartridges are owned by either a central manager (tape library
manager) or a peer-to-peer management implementation. These devices are dynamically
allocated and reallocated to systems (tape library clients) as required, based on demand.
Tape pooling allows for resource sharing, automation, improved tape management, and
added security for tape media.
217
Software is required to manage the assignment and locking of the tape devices to serialize
tape access. Tape pooling is an efficient and cost-effective way of sharing expensive tape
resources, such as automated tape libraries. At any particular instant in time, a tape drive can
be owned by one system only.
This concept of tape resource sharing and pooling is proven in medium-to-enterprise backup
and archive solutions that use, for example, IBM Tivoli Storage Manager with SAN-attached
IBM tape libraries.
Logical disk volumes are defined within the storage subsystem and assigned to servers. The
logical disk is addressable from the server. A logical disk might be a subset or superset of
disks that are only addressable by the subsystem itself. A logical disk volume can also be
defined as subsets of several physical disks (striping). The capacity of a disk volume is set
when the logical disk is defined.
For example, two logical disks, with different capacities (for example, 50 GB and 150 GB)
might be created from a single 300 GB hardware-addressable disk. Each of the two disks is
assigned to a different server, which leaves 100 GB of unassigned capacity. A single 2,000
GB logical disk might also be created from multiple real disks that exist in different storage
subsystems. The underlying storage controller must have the necessary logic to manage the
volume grouping and guarantee access securely to the data.
The function of a storage controller can be further used by certain storage virtualization
engines, such as the IBM SAN Volume Controller. This engine, when compared to
environments that do not use this controller, offers better and more scalability and
virtualization of storage resources. The SAN Volume Controller provides these benefits with
less management effort and clearer visibility to the target host systems.
218
Figure 10-4 shows multiple servers that are accessing logical volumes that were created by
using the different alternatives that we mentioned. (The logical volume, which is called the
unallocated volume, is not assigned to any server.)
Linux
Windows
AIX
z/OS
Logical Volume 3
Unallocated volume
LV 1
LV 2
LV 4
LV 5
Logical Volume 6
10.3.3 Consolidation
We can improve scalability, security, and manageability by enabling devices in separate SAN
fabrics to communicate without merging fabrics into a single, large SAN fabric. This capability
enables clients to initially deploy separate SAN solutions at the departmental and data center
levels and then to consolidate them into large enterprise SAN solutions. This consolidation
occurs as clients experience and requirements grow and change. This type of solution is also
known as data center bridging.
Clients deploy multiple SAN islands for different applications with different fabric switch
solutions. The growing availability of iSCSI server capabilities creates the opportunity for
low-cost iSCSI server integration and storage consolidation. Additionally, depending on the
choice of router, iSCSI servers can provide Fibre Channel over IP (FCIP) or Internet Fibre
Channel Protocol (iFCP) capability.
The available multiprotocol SAN routers provide an iSCSI Gateway Service to integrate
low-cost Ethernet-connected servers to existing SAN infrastructures. The iSCSI Gateway
Service also provides a Fibre Channel-to-Fibre Channel (FC-FC) Routing Service to
interconnect multiple SAN islands without requiring the fabrics to merge into a single large
SAN.
219
Figure 10-5 shows an example of using a multiprotocol router and converged core switch to
extend SAN capabilities across long distances or merely over metropolitan areas.
NAS
iSCSI
iSCSI
FC
FC
DCB/CEE
FCIP
FC
Multiprotocol router
FC
FCoE
Downtime
Purchase of more switches and ports
Purchase of HBAs
Migration costs
Configuration costs
Purchase of more licenses
Ongoing maintenance
220
The router and core switch can provide more benefits. In this example, which is an FC-FC
routing service that negates the need for a costly SAN fabric merger, the advantages are
apparent and real. A router can also provide the following benefits:
The following list provides a summary of the key benefits of upgrading to a converged
network:
221
222
2. Fabric convergence.
The second step is to implement more core types of switches that support data center
bridging and converged network protocols. Therefore, rather than implementing a
converged network on TOR switches or blades, we move this function to the core
directors. A second stage of the development of DCB standards introduces multi-hop
bridging because different solutions are available from each of the vendors of the SAN
networking products (Figure 10-8).
223
3. Storage convergence.
For the final step of the migration, we implement native FCoE-enabled storage devices.
Now, various vendors with midrange to enterprise disk storage systems already offer
FCoE. This step enables clients to migrate the current FC-attached storage data to the
FCoE-enabled storage system and disconnect the original FC core and edge switches.
This step dramatically reduces the requirements for operation and management of the
infrastructure, reduces the power consumption, and simplifies the complexity of the
network (rack space and cabling). Figure 10-9 shows the final status of the converged
network infrastructure.
FCoE in the converged network offers several benefits and advantages over existing
approaches to I/O consolidation:
Compatibility with existing Fibre Channel SANs by preserving well-known Fibre
Channel concepts. Examples of these concepts include virtual SANs (VSANs),
worldwide names (WWNs), FC IDs (FCIDs), multipathing, and zoning to servers and
storage arrays.
A high level of performance, which is comparable to the performance of current
Ethernet and Fibre Channel networks. These networks are achieved by using a
hardware-based Ethernet network infrastructure that is not limited by the overhead of
higher-layer TCP/IP protocols.
The exceptional scalability of Ethernet at the highest available speeds (1 GbE, 10 GbE,
and 40 GbE, and eventually 100 GbE).
Simplified operations and management (no change to the management infrastructure
that is deployed in the SANs).
224
225
In Figure 10-10, an application runs on system AIX2. The system manages mirrored disks
from both of the storage systems (green and orange). SAN zoning allows both cluster nodes
(AIX2 and AIX3) to operate the same set of disks. The cluster has one primary cluster node
that is active (an active-passive cluster). When AIX2 fails, cluster services that run on AIX3
recognize the failure and automatically move all of the application resources to AIX3. The disk
sets are activated on AIX3, and the application is started in the correct sequence.
226
SAN zoned disks are available to all four AIX host systems. The clusters are active, and the
master application works concurrently on them with workload balancing. If the AIX2 system
fails, the cluster application automatically deactivates the assigned disks and redistributes the
workload among the remaining active cluster nodes. No interruption occurs to the business.
This configuration is costly and typically only employed for business critical applications, such
as banking systems and air traffic control systems.
227
In the tape drive and tape library sharing approach, the backup and recovery server or client
that requests that a backup copy is copied to or from tape reads or writes the data directly to
the tape device by using SCSI commands. This approach bypasses the network transports
latency and network protocol path length. Therefore, it can offer improved backup and
recovery speeds in cases where the network is the constraining factor.
The data is read from the source device and written directly to the destination device. The
central backup and recovery server controls only the tape mount operations and stores the
references (metadata) in its embedded database system.
Figure 10-12 shows an example of tape library sharing and LAN-free backups.
IBM Lotus Domino and IBM DB2 database systems benefit from the improved performance
of backups directly to tapes over Fibre Channel. However, small servers with a high number
of files still continue to back up to a LAN or WAN.
IBM offers enterprises a centralized backup and recovery solution that supports various
platforms and database systems. IBM Tivoli Storage Manager and its component IBM Tivoli
Storage Manager for SANs enable clients to perform online backups and archives of large
application systems directly to tape over a SAN without a significant effect on performance.
228
You can now perform extended-distance backups for disaster recovery within a campus or
city (Figure 10-13).
Standby Server
Primary Server
SAN
primary site
Primary
Storage
SAN Bridge
(FC, FCIP, FCoE)
Primary
Tape Library
SAN
disaster site
Tape Library
Copypool
Storage
Replication
When longer distances are required, SANs must be connected by using gateways and WANs.
One of the solutions is FCoE.
Depending on your business requirements, disaster protection deployments might use copy
services that are implemented in disk subsystems and tape libraries (that might be achieved
by using SAN services), SAN copy services, or most likely a combination of both.
229
230
An IBM ILM solution in a tiered storage environment is designed with the following factors:
Reduces the TCO of managing information. It can help optimize data costs and
management by freeing expensive disk storage for the most valuable information.
Segments data according to value. Segmenting data can help to create an economical
balance and sustainable strategy to align storage costs with business objectives and
information value.
Helps you decide about moving, retaining, and deleting data because ILM solutions are
closely tied to applications.
Manages information and determines how the information is managed based on content,
rather than migrating data that is based on technical specifications. This approach can
help result in more responsive management. This solution offers you the ability to retain or
delete information according to your business rules.
Provides the framework for a comprehensive enterprise content management strategy.
231
Email
Instant messages
Business transactions
Accounting records
Contracts
Insurance claims processing
All of these types of information can have different retention periods. These periods can be
two years, seven years, or forever (permanent retention). Data is an asset when it must be
kept; however, data that is kept past its mandated retention period might also become a
liability. Furthermore, the retention period can change because of factors, such as litigation.
All of these factors mandate tight coordination and the need for ILM.
In addition to the numerous state and governmental regulations that must be met for data
storage, industry-specific and company-specific regulations also must be met. And these
regulations are constantly updated and amended. Organizations must develop a strategy to
ensure that the correct information is kept for the correct period of time, and that the
information is readily accessible when it must be retrieved at the request of regulators or
auditors.
It is easy to envision the exponential growth in data storage that results from these
regulations and the accompanying requirement for a means of managing this data. Overall,
the management and control of retention-managed data is a significant challenge for the IT
industry when you consider factors, such as cost, latency, bandwidth, integration, security,
and privacy.
232
ILM can potentially provide the framework for a comprehensive information management
strategy and help ensure that information is stored on the most cost-effective media. This
framework helps enable administrators to use tiered and virtual storage, and to process
automation. By migrating unused data off more costly, high-performance disks, ILM can help
by performing the following functions:
ILM recognizes that different types of information have different values at different points in
their lifecycles.
Data can be allocated to a specific storage level that is aligned to its cost, with policies that
define when and where data is moved (Figure 10-16).
Data expired or
moved automatically
to lower-cost storage.
Data becomes inactive.
Automated movement
by policy enforcement.
233
234
11
Chapter 11.
235
Data center
IT Resources
IT Load
Processor
55%
45%
70%
30%
Power supply,
memory, fans, planar,
drives . . .
Utilization
Resource
usage rate
80%
20%
Idle
In typical data centers, the IT equipment does not use 55% of the overall energy that is
brought into the data center. Therefore, this portion of the energy is not producing
calculations, data storage, and so on. The concept of a green data center is to eliminate this
waste and reduce such inefficiency.
Energy conversion: Basic laws of thermodynamics state that energy cannot be created or
destroyed; it changes only in form. The efficiency of this conversion is less than 100% (in a
real world, much less than 100%).
Solution designers and IT architects must also consider the energy consumption of the
components at the IT equipment level. For example, in a typical server, the processor uses
only 30% of the energy and the remainder of the system uses 70%. Therefore, efficient
hardware design is crucial. Features, such as the virtualization of physical servers, can help
to change this ratio to a more reasonable value.
237
Companies need to consider the use of IT resources in the data center. A typical server
utilization rate is around 20%. Underutilized systems can be a significant issue because
significant energy is expended on non-business activities, therefore wasting a major
investment. Again, server virtualization, consolidation, and the addressed provisioning of IT
resources help to use the entire capacity of your IT equipment.
Data centers must become immensely more efficient to meet their needs while they keep
costs in check as the demand for and price of resources continue to rise. But the realization of
this efficiency requires a deep and pervasive transformation in how data centers are
designed, managed, operated, populated, and billed. These aspects mandate a unified and
coordinated effort across organizational and functional boundaries toward a common set of
goals.
We introduce the concept of green data centers and how IBM supports the migration to
next-generation data centers that are effective, cost-efficient, and environment friendly.
238
Physical consolidation:
Consolidate many servers into fewer servers on physical resource boundaries
Reduce system management complexity
Reduce the physical footprint of servers in the data center
Virtualization:
Application integration:
239
The storage strategy for green data centers includes the following elements:
Slow/Low
100+ Years
50 Years
20 Years
5 Years
3 Years
2 Months
1 Day
1 Hour
However, infrequently accessed or inactive data can become suddenly valuable again as
events occur, or as new business initiatives or projects are taken on. Historically, the
requirement to retain information results in a buy more storage mentality. However, this
approach increases overall operational costs, complexity, and the demand for hard-to-find
qualified personnel.
240
Typically, only around 20% of the information is active and frequently accessed by users. The
remaining 80% is either inactive or even obsolete (Figure 11-3).
241
This logical view can hide undesirable characteristics of storage while it presents storage in a
more convenient manner for applications. For example, storage virtualization might present
storage capacity as a consolidated whole, hiding the actual physical boxes that contain the
storage.
In this way, storage becomes a logical pool of resources that exist virtually, regardless of
where the actual physical storage resources are in the larger information infrastructure.
These software-defined virtual resources are easier and less disruptive to change and
manage than hardware-based physical storage devices because they do not involve moving
equipment or making physical connections. As a result, they can respond more flexibly and
dynamically to changing business needs. Similarly, the flexibility that is afforded by virtual
resources makes it easier to match storage to business requirements.
242
The value of a virtualized infrastructure is in the increased flexibility that is created by having
pools of system resources on which to draw and in the improved access to information that is
afforded by a shared infrastructure. Value is also a result of the lower total cost of ownership
(TCO) that comes with decreased management costs, increased asset utilization, and the
ability to link infrastructure performance to specific business goals.
For more information about how IBM Storage Virtualization solutions can help your
organization meet its storage challenges, see the IBM Information Infrastructure Solutions
Handbook, SG24-7814, or see this website:
http://www.ibm.com/systems/storage/virtualization/
End-to-end storage provisioning is the term that is applied to the whole set of steps that are
required to provision usable storage capacity to a server. Provisioning covers the
configuration of all of the elements in the chain. This process includes the steps from carving
out a new volume on a storage subsystem, through creating a file system at the host and
making it available to the users or applications.
Typically, this process involves a storage administrator that uses several different tools, each
focused on a specific task, or the tasks are spread across several people. This spreading of
tasks and tools creates many inefficiencies in the provisioning process, which affect the ability
of IT departments to respond quickly to changing business demands. The resulting
complexity and high degree of coordination can also lead to errors and can possibly affect the
systems and application availability.
Automation of the end-to-end storage provisioning process by using workflow automation can
significantly simplify this task of provisioning storage capacity. Each step is automated and
the rules for preferred practices around zoning, device configuration, and path selection can
be applied automatically. The benefits are increased responsiveness to business
requirements, lower administrative costs, and higher application availability.
243
In a tiered storage environment, data is classified and assigned dynamically to different tiers.
For example, we can use expensive, fast-performance storage components to store
often-accessed and mission-critical files, in contrast with using less expensive storage for less
frequently used non-critical data. Tiered storage improves efficiency and saves costs. We can
identify the following typical storage tiers, which are categorized based on performance and
cost for each gigabyte:
High-performance SAN-attached disk systems (solid-state drive (SSD) or serial-attached
Small Computer System Interface (SCSI) (SAS))
Medium-performance SAN-attached disks (SAS or Serial Advanced Technology
Attachment (SATA))
Network-attached storage (NAS) systems
Tape storage and other media with sequential access
Each level of storage tier can be assigned manually by a storage administrator, or data can be
moved automatically between tiers, which is based on migration policies. Figure 11-5 shows
the conceptual model of storage tiering.
IBM offers various tools and utilities for storage tiering and hierarchical management for
different scenarios. Tools range from IBM Easy Tier, which is used in enterprise disk storage
systems, up through IBM Tivoli Storage Manager for Hierarchical Storage Management for
Windows and IBM Tivoli Storage Manager for Space Management for the AIX and Linux
platform. For tiered management, IBM offers IBM Global Parallel File System (GPFS) for
data migration between different levels of storage.
244
Compression immediately reduces the amount of required physical storage across all storage
tiers. This solution, which supports the online compression of existing data, allows storage
administrators to gain back free disk space in the existing storage system. You can compress
data without changing administrative processes or forcing users to clean up or archive data.
The benefits to the business are immediate because the capital expense of upgrading the
storage system is delayed. Because data is stored in compressed format in the primary
storage system, all other storage tiers and the transports in between realize the same
benefits. Replicas, backup images, and replication links all require fewer expenditures after
the implementation of compression at the source.
After compression is applied to the stored data, the required power and cooling for each unit
of storage are reduced. This reduction is possible because more logical data is stored on the
same amount of physical storage. In addition, within a particular storage system, more data
can be stored; therefore, fewer overall rack units are required. Figure 11-6 on page 246
shows the typical compression rates that can be achieved with specific IBM products.
The exact compression ratio depends on the nature of the data. IBM documented
compression ratios as high as 90% in certain Oracle database configurations and about 50%
with PDF files. As always, compression ratios vary by data type and how the data is used.
In contrast to compression, the data deduplication mechanism identifies identical chunks of
data within a storage container. This process keeps only one copy of each chunk. All of the
other logically identical chunks point to this chunk. Various implementations of this method
exist. One option is inline data deduplication and the other option is post-processing data
deduplication. Each chunk of data must be identified in a way that is easily comparable.
Chunks are processed by using a parity calculation or cryptographic hash function.
This processing gives the chunks shorter identifiers that are known as hash values, digital
signatures, or fingerprints. These fingerprints can be stored in an index or catalog where they
can be compared quickly with other fingerprints to find matching chunks.
245
Figure 11-6 shows the typical compression rates that can be achieved with specific IBM
products.
Data deduplication processing can occur on the client, an infrastructure server, or the storage
system. Each option has factors to consider:
Client-based data deduplication
This process reduces the amount of data that is transferred over the network to the
storage system. But, this option can require extra CPU and disk I/O processing on the
client side.
Server-based data deduplication
This process deduplicates the data of multiple clients at a scheduled time. But, this
process requires extra CPU and disk I/O processing on the infrastructure server, for
example, the IBM Tivoli Storage Manager server.
Storage-based data deduplication
This process occurs at the disk storage device level, where the data is stored. This type of
data deduplication is generally transparent to the clients and servers. This process uses
CPU and disk I/O on the storage system.
For more information about data compression, data deduplication, and concrete solutions
from IBM, see Introduction to IBM Real-time Compression Appliances, SG24-7953, and
Implementing IBM Storage Data Deduplication Solutions, SG24-7888.
246
12
Chapter 12.
247
For more information about other IBM storage products, see the IBM System Storage
Solutions Handbook, SG24-5250.
For more information about each product and its market position, see this IBM storage
website:
http://www.ibm.com/systems/storage/
248
For more information about the latest IBM SAN products, see this website:
http://www.ibm.com/systems/networking/switches/san/index.html
Chapter 12. IBM Fibre Channel storage area network product portfolio
249
Figure 12-2 Front view of the IBM System Storage SAN24B-4 Express switch
A single SAN24B-4 Express switch can serve as the cornerstone of a SAN for individuals that
want to obtain the benefits of storage consolidation and implement Fibre Channel storage
systems. This entry-level configuration can consist of one or two Fibre Channel links to a disk
storage array or to a Linear Tape Open (LTO) tape drive. An entry-level, 8-port storage
consolidation solution can support up to seven servers with a single path to disk or tape.
The Ports on Demand (PoD) feature enables a base switch to grow to 16 ports and 24 ports
to support more servers and more storage devices without taking the switch offline. A high
availability (HA) solution can be created with redundant switches. This capability is ideal for
server clustering environments.
This configuration can support 6 - 22 servers, each with dual Fibre Channel adapters that are
cross-connected to redundant SAN24B-4 Express switches. These switches are
cross-connected to a dual-controller storage system. The focus of the SAN24B-4 Express is
as the foundation of small to medium-sized SANs.
However, the SAN24B-4 Express can be configured to participate as a full member in an
extended fabric configuration with other members of the IBM System Storage and former
TotalStorage SAN b-type and m-type families. This capability helps provide investment
protection as SAN requirements evolve and grow.
250
Adaptive Networking
Advance Performance Monitor
Extended Fabric
Fabric Watch
Trunking Activation
Server Application Optimization (SAO)
For more information about entry SAN switches, see this website:
http://www.ibm.com/systems/storage/san/entry/index.html
Chapter 12. IBM Fibre Channel storage area network product portfolio
251
Powered by Cisco NX-OS Software and Cisco Data Center Network Manager (DCNM)
software, the Cisco MDS 9396S delivers advanced storage networking features and functions
that combine with ease of management and compatibility with the entire Cisco MDS 9000
Family portfolio for reliable end-to-end connectivity. Figure 12-3 shows the Cisco MDS 9396S.
252
Chapter 12. IBM Fibre Channel storage area network product portfolio
253
254
Diagnostic port (D_port) feature for physical media diagnostic, troubleshooting, and
verification services
Multi-tenancy in cloud environments through Virtual Fabrics, Integrated Routing, QoS, and
fabric-based zoning features
12.4.4 Cisco MDS 9148S 16G Multilayer Fabric Switch for IBM System Storage
The Cisco MDS 9148S 16G Multilayer Fabric Switch for IBM System Storage (Figure 12-6) is
the latest generation of the highly reliable, flexible, and low-cost Cisco MDS 9100 Series
switches. It combines high performance with exceptional flexibility and cost-effectiveness.
This powerful, compact one rack-unit (1RU) switch scales from 12 to 48 line-rate 16 Gbps
Fibre Channel ports.
The Cisco MDS 9148S is excellent for the following functions:
A stand-alone SAN in small departmental storage environments
A top-of-the-rack switch in medium-sized redundant fabrics
An edge switch in enterprise data center core-edge topologies
Chapter 12. IBM Fibre Channel storage area network product portfolio
255
256
This switch enables the transparent integration of new technologies into the data center to
provide highly flexible SAN solutions (see Figure 12-7).
257
Provides deterministic hardware performance and features so that virtual machines (VMs)
include the same SAN attributes as physical servers
Enables outstanding levels of availability and reliability
Ships mainframe-ready, with full support for IBM z Systems FICON and Fibre Channel
Protocol (FCP) environments
Includes modular, multilayer, highly available, and dual-supervisor modules with six fabric
slots and eight module slots (14RU)
Offers 384 two-Gbps, 4-Gbps, 8-Gbps, 16-Gbps, and 10-Gbps full line-rate autosensing
Fibre Channel ports in a single chassis for deployment in both open systems and FICON
environments
Includes 384 ten-Gbps full line-rate autosensing FCoE ports in a single chassis
Includes a 48-port 16-Gbps Fibre Channel switching module
Features a 48-port 10-Gbps FCoE module
Includes hot-swappable switching modules, supervisor modules, fans, power supplies,
and small form-factor pluggables (SFPs)
Provides front-to-back airflow
258
Integrated ISL data compression and encryption offer bandwidth optimization and data
protection.
Compatibility with existing infrastructure minimizes the need to rip out and replace
equipment.
Low overhead and low latency eliminate I/O bottlenecks and unleash the full performance
of Flash, SSD, and 16 Gbps-capable storage.
In addition, Brocade Gen 5 Fibre Channel platforms offer breakthrough technologies that
dramatically simplify SAN deployment and management and drive down operational costs:
UltraScale chassis connectivity enables higher density and simpler fabrics that reduce
network complexity and cost.
Fabric Vision technology maximizes uptime, optimizes application performance, and
simplifies SAN management through innovative diagnostic, monitoring, and management
technology.
The SAN768B-2 and SAN384B-2 are efficient at reducing power consumption, cooling, and
the carbon footprint in data centers. Although these switches provide exceptional
performance and scale, these networking backbones use less than 0.2 watts/Gbps.
As members of the IBM System Storage family of b-type SAN products, the SAN768B-2 and
the SAN384B-2 participate in fabrics that contain other b-type and m-type devices that are
manufactured by Brocade. This versatile hardware can serve as the backbone in a complex
fabric and provide connections to other b-type and m-type directors, switches, and routers.
Figure 12-8 shows the directors.
Figure 12-8 IBM System Storage SAN384B-2 (left) and SAN768B-2 (right)
Chapter 12. IBM Fibre Channel storage area network product portfolio
259
260
Auditing
Syslog
Command-line interface (CLI)
Browser-based Web Tools
IBM Network Advisor
Full compatibility with an earlier version with IBM System Storage and IBM TotalStorage
b-type and m-type SAN directors, switches, and routers; other directors, switches, and
routers that are manufactured by Brocade.
Chapter 12. IBM Fibre Channel storage area network product portfolio
261
12.5.3 Cisco MDS 9706 Multilayer Director for IBM System Storage
The Cisco MDS 9706 Multilayer Director for IBM System Storage is a director-class SAN
switch for deployment in small to mid-sized storage networks that can support enterprise
clouds and business transformation. The switch layers a comprehensive set of intelligent
features onto a high-performance, protocol-independent switch fabric.
The MDS 9706 (Figure 12-9) addresses the stringent requirements of large virtualized data
center storage environments. It delivers uncompromising availability, security, scalability, ease
of management, and transparent integration of new technologies for flexible data center SAN
solutions. It shares the operating system and management interface with other Cisco data
center switches. By using the Cisco MDS 9706, you can transparently deploy unified fabrics
with Fibre Channel, FICON, and FCoE connectivity for low TOC.
Chapter 12. IBM Fibre Channel storage area network product portfolio
263
Each Supervisor-2 Module provides the necessary crossbar bandwidth to deliver full system
performance in the MDS 9506 director with up to four Fibre Channel switching modules. It
provides that loss or removal of a single crossbar module that has no effect on system
performance. The Fibre Channel switching modules optimize performance, flexibility, and
density. The Cisco MDS 9506 Multilayer Director requires a minimum of one and allows a
maximum of four switching modules.
These modules are available in 12-port, 24-port, and 48-port 4 Gbps configurations, which
allow the Cisco MDS 9506 to support 12 - 192 Fibre Channel ports for each chassis.
Optionally, a 4-port 10 Gbps Fibre Channel module is available for high-performance ISL
connections over metropolitan optical networks.
Advanced traffic management capabilities are integrated into the switching modules to help
simplify deployment and to optimize performance across a large fabric. The PortChannel
capability allows users to aggregate up to 16 physical 2 Gbps inter-switch links into a single
logical bundle, which provides optimized bandwidth utilization across all links. The bundle
might span any port from any 16-port switching module within the chassis, which provides up
to 32 Gbps throughput.
The following switching modules are available:
264
Figure 12-11 shows the front view of the Cisco MDS 9513 chassis.
The Cisco MDS 9513 Multilayer Director uses two Supervisor-2 Modules, which support high
availability. The Supervisor-2 Module provides industry-leading scalability, intelligent SAN
services, nondisruptive software upgrades, graceful process restart and failover, and
redundant operation. Dual crossbar-switching fabric modules provide a total internal switching
bandwidth of 2.4 Tbps for inter-connection of up to 11 Fibre Channel switching modules.
Fibre Channel switching modules improve performance, flexibility, and density. The Cisco
MDS 9513 for IBM System Storage requires a minimum of one Fibre Channel switching
module and allows a maximum of 11. These modules are available in 12-port, 24-port, or
48-port 4 Gbps and 8 Gbps configurations, which enable the Cisco MDS 9513 to support
12 - 528 Fibre Channel ports for each chassis. Optionally, a 4-port 10 Gbps Fibre Channel
module is available for high-performance ISL connections over metropolitan optical networks.
The Cisco MDS 9513 includes the following features:
Supports Fibre Channel throughput of up to 8 Gbps for each port and up to 64 Gbps with
each PortChannel ISL connection
Offers Gigabit Ethernet (GbE) IP ports for iSCSI or FCIP connectivity over global networks
Offers scalability for 12 - 528 one-Gbps, 2-Gbps, 4-Gbps, and 8-Gbps Fibre Channel ports
Includes a high availability design with support for nondisruptive firmware upgrades
Includes VSAN capability for SAN consolidation into virtual SAN islands on a single
physical fabric
Supports Enterprise, SAN Extension over IP, Mainframe, Storage Services Enabler, and
Fabric Manager Server Packages to provide added intelligence and value
For more information, see this website:
http://www.ibm.com/systems/storage/san/enterprise/
Chapter 12. IBM Fibre Channel storage area network product portfolio
265
Chapter 12. IBM Fibre Channel storage area network product portfolio
267
268
Chapter 12. IBM Fibre Channel storage area network product portfolio
269
270
13
Chapter 13.
Certification
In this chapter, we provide an insight into several professional certifications that relate to the
topics in this book.
271
272
273
274
Related publications
The publications that are listed in this section are considered particularly suitable for a more
detailed discussion of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide additional information about the topic in this
document. Note that some publications referenced in this list might be available in softcopy
only:
IBM TotalStorage: SAN Product, Design, and Optimization Guide, SG24-6384
Implementing an Open IBM SAN, SG24-6116
Implementing the Cisco MDS 9000 in an Intermix FCP, FCIP, and FICON Environment,
SG24-6397
Introduction to SAN Distance Solutions, SG24-6408
Introducing Hosts to the SAN Fabric, SG24-6411
IP Storage Networking: IBM NAS and iSCSI Solutions, SG24-6240
The IBM TotalStorage NAS Integration Guide, SG24-6505
Implementing the IBM TotalStorage NAS 300G: High Speed Cross Platform Storage and
Tivoli SANergy!, SG24-6278
Using iSCSI Solutions Planning and Implementation, SG24-6291
IBM Storage Solutions for Server Consolidation, SG24-5355
Implementing the Enterprise Storage Server in Your Environment, SG24-5420
Implementing Linux with IBM Disk Storage, SG24-6261
IBM Tape Solutions for Storage Area Networks and FICON, SG24-5474
IBM Enterprise Storage Server, SG24-5465
IBM System Storage Solutions Handbook, SG24-5250
IBM Storage Infrastructure for Business Continuity, REDP-4605
IBM System Storage Data Encryption, SG24-7797
Implementing the Storwize V7000 and the IBM System Storage SAN32B-E4 Encryption
Switch, SG24-7977
ILM Library: Information Lifecycle Management Best Practices Guide, SG24-7251
IBM Information Infrastructure Solutions Handbook, SG24-7814
Introduction to IBM Real-time Compression Appliances, SG24-7953
Implementing IBM Storage Data Deduplication Solutions, SG24-7888
You can search for, view, download or order these documents and other Redbooks,
Redpapers, Web Docs, draft and additional materials, at the following website:
http://www.redbooks.ibm.com/
275
Online resources
The following websites are also relevant as further information sources:
IBM Storage hardware, software, and solutions:
http://www.storage.ibm.com
IBM System Storage storage area networks:
http://www.ibm.com/systems/storage/san/
Brocade:
http://www.brocade.com
Cisco:
http://www.cisco.com
QLogic:
http://www.qlogic.com
Emulex:
http://www.emulex.com
Finisar:
http://www.finisar.com
IBM Tivoli software:
http://www.ibm.com/software/tivoli
IEEE:
http://www.ieee.org
Storage Networking Industry Association:
http://www.snia.org
Fibre Channel Industry Association:
http://www.fibrechannel.com
SCSI Trade Association:
http://www.scsita.org
Internet Engineering Task Force:
http://www.ietf.org
American National Standards Institute:
http://www.ansi.org
Technical Committee T10:
http://www.t10.org
Technical Committee T11:
http://www.t11.org
276
Related publications
277
278
SG24-5470-06
ISBN 0738441430
(0.5 spine)
0.475<->0.873
250 <-> 459 pages
Back cover
SG24-5470-06
ISBN 0738441430
Printed in U.S.A.
ibm.com/redbooks