sg248333
sg248333
sg248333
Hiroyuki Miyoshi
Khanh Ngo
Arnold Byron Lua
Yuka Sasaki
Yasuhiro Yoshihara
Larry Coyne
Redbooks
International Technical Support Organization
March 2022
SG24-8333-09
Note: Before using this information and the product it supports, read the information in “Notices” on
page ix.
This edition applies to Version 1, Release 3, Modification 2, Fix Level 2 of IBM Spectrum Archive Enterprise
Edition (product number 5639-LP1).
© Copyright International Business Machines Corporation 2015, 2020, 2022. All rights reserved.
Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule
Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .x
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv
Summary of changes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
December 2021, Tenth Edition Version 1.3.2.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
June 2021, Ninth Edition Version 1.3.1.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
February 2020, Eighth Edition Version 1.3.0.6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
April 2019, Seventh Edition Version 1.3.0 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
June 2018, Sixth Edition Version 1.2.6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
January 2018, Fifth Edition Version 1.2.5.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
August 2017, Fourth Edition Version 1.2.4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
February 2017, Third Edition Version 1.2.2 minor update . . . . . . . . . . . . . . . . . . . . . . . . . . xix
January 2017, Third Edition Version 1.2.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix
August 2016, Second Edition Version 1.2.1. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx
June 2016, First Edition, Version 1.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx
January 2015, Third Edition Version 1.1.1.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxi
November 2014, Second Edition Version 1.1.1.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxii
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. iii
2.2.5 Active File Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
2.2.6 Scale Out Backup and Restore. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.3 OpenStack SwiftHLM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
2.4 IBM Spectrum Archive EE dashboard. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
2.5 IBM Spectrum Archive EE REST API . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
2.6 Types of archiving . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
iv IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
5.1.1 Configuration worksheet tables. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
5.1.2 Obtaining configuration information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
5.1.3 Configuring key-based login with OpenSSH . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
5.1.4 Preparing the IBM Spectrum Scale file system for IBM Spectrum Archive EE . . 104
5.2 Configuring IBM Spectrum Archive EE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.2.1 The ltfsee_config utility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.2.2 Configuring a single node cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
5.2.3 Configuring a multiple-node cluster. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
5.2.4 Configuring a multiple-node cluster with two tape libraries . . . . . . . . . . . . . . . . . 115
5.2.5 Modifying a multiple-node configuration for control node redundancy . . . . . . . . 116
5.3 First-time start of IBM Spectrum Archive EE. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
5.3.1 Configuring IBM Spectrum Archive EE with IBM Spectrum Scale AFM . . . . . . . 119
5.3.2 Configuring a Centralized Archive Repository solution . . . . . . . . . . . . . . . . . . . . 120
5.3.3 Configuring an Asynchronous Archive Replication solution . . . . . . . . . . . . . . . . 123
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 129
6.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
6.1.1 IBM Spectrum Archive EE command summaries . . . . . . . . . . . . . . . . . . . . . . . . 131
6.1.2 Using the command-line interface. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6.2 Status information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
6.2.1 IBM Spectrum Scale . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
6.2.2 IBM Spectrum Archive Library Edition component . . . . . . . . . . . . . . . . . . . . . . . 137
6.2.3 Hierarchical Space Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
6.2.4 IBM Spectrum Archive EE. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
6.3 Upgrading components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
6.3.1 IBM Spectrum Scale . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
6.3.2 IBM Spectrum Archive LE component . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
6.3.3 Hierarchical Storage Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
6.3.4 IBM Spectrum Archive EE. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
6.4 Starting and stopping IBM Spectrum Archive EE . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
6.4.1 Starting IBM Spectrum Archive EE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
6.4.2 Stopping IBM Spectrum Archive EE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
6.5 Task command summaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
6.5.1 eeadm task list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
6.5.2 eeadm task show . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
6.6 IBM Spectrum Archive EE database backup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
6.7 IBM Spectrum Archive EE automatic node failover . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6.7.1 IBM Spectrum Archive EE monitoring daemon. . . . . . . . . . . . . . . . . . . . . . . . . . 153
6.8 Tape library management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
6.8.1 Adding tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
6.8.2 Moving tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6.8.3 Formatting tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158
6.8.4 Removing tape drives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160
6.8.5 Adding tape drives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
6.9 Tape storage pool management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
6.9.1 Creating tape cartridge pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
6.9.2 Deleting tape cartridge pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
6.10 Pool capacity monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
6.11 Migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
6.11.1 Managing file migration pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168
6.11.2 Threshold-based migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168
6.11.3 Manual migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173
6.11.4 Replicas and redundant copies. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
Contents v
6.11.5 Data Migration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
6.11.6 Migration hints and tips . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
6.12 Premigration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184
6.12.1 Premigration with the eeadm premigrate command . . . . . . . . . . . . . . . . . . . . . 185
6.12.2 Premigration running the mmapplypolicy command . . . . . . . . . . . . . . . . . . . . . 185
6.13 Preserving file system objects on tape . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186
6.13.1 Saving file system objects with the eeadm save command . . . . . . . . . . . . . . . 186
6.13.2 Saving file system objects with policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
6.14 Recall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
6.14.1 Transparent recall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
6.14.2 Selective recall using the eeadm recall command . . . . . . . . . . . . . . . . . . . . . . 191
6.14.3 Read Starts Recalls: Early trigger for recalling a migrated file . . . . . . . . . . . . . 192
6.14.4 Recommend Access Order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193
6.15 Recalling files to their resident state . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196
6.16 Reconciliation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
6.17 Reclamation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
6.17.1 Reclamation considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
6.18 Checking and repairing tapes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
6.19 Importing and exporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
6.19.1 Importing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
6.19.2 Exporting tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
6.19.3 Offlining tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
6.20 Drive Role settings for task assignment control . . . . . . . . . . . . . . . . . . . . . . . . . . . . 207
6.21 Tape drive intermix support. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208
6.21.1 Objective for WORM tape support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209
6.21.2 Function overview for WORM tape support . . . . . . . . . . . . . . . . . . . . . . . . . . . 210
6.21.3 The effects of file operations on immutable and appendOnly files . . . . . . . . . . 211
6.22 Obtaining the location of files and data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213
6.23 Obtaining system resources, and tasks information . . . . . . . . . . . . . . . . . . . . . . . . . 214
6.24 Monitoring the system with SNMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
6.25 Configuring Net-SNMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 218
6.25.1 Starting and stopping the snmpd daemon . . . . . . . . . . . . . . . . . . . . . . . . . . . . 219
6.25.2 Example of an SNMP trap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 219
6.26 IBM Spectrum Archive REST API . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 219
6.26.1 Pools endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 220
6.26.2 Tapes endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223
6.26.3 Libraries endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 225
6.26.4 Nodegroups endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 226
6.26.5 Nodes endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 227
6.26.6 Drives endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 228
6.26.7 Task endpoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
6.27 File system migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231
vi IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.10 Use cases for mmapplypolicy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244
7.10.1 Creating a traditional archive system policy . . . . . . . . . . . . . . . . . . . . . . . . . . . 244
7.10.2 Creating active archive system policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
7.10.3 IBM Spectrum Archive EE migration policy with AFM. . . . . . . . . . . . . . . . . . . . 246
7.11 Capturing a core file on Red Hat Enterprise Linux with the Automatic Bug Reporting Tool
247
7.12 Anti-virus considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248
7.13 Automatic email notification with rsyslog. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248
7.14 Overlapping IBM Spectrum Scale policy rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248
7.15 Storage pool assignment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
7.16 Tape cartridge removal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
7.16.1 Reclaiming tape cartridges before you remove or export them . . . . . . . . . . . . 251
7.16.2 Exporting tape cartridges before physically removing them from the library. . . 251
7.17 Reusing LTFS formatted tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
7.17.1 Reformatting LTFS tape cartridges through eeadm commands . . . . . . . . . . . . 252
7.18 Reusing non-LTFS tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
7.19 Moving tape cartridges between pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
7.19.1 Avoiding changing assignments for tape cartridges that contain files. . . . . . . . 254
7.19.2 Reclaiming a tape cartridge and changing its assignment . . . . . . . . . . . . . . . . 254
7.20 Offline tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
7.20.1 Do not modify the files of offline tape cartridges . . . . . . . . . . . . . . . . . . . . . . . . 254
7.20.2 Solving problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
7.21 Scheduling reconciliation and reclamation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
7.22 License Expiration Handling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
7.23 Disaster recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
7.23.1 Tiers of disaster recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
7.23.2 Preparing IBM Spectrum Archive EE for a tier 1 disaster recovery strategy (offsite
vaulting) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
7.23.3 IBM Spectrum Archive EE tier 1 DR procedure . . . . . . . . . . . . . . . . . . . . . . . . 259
7.24 IBM Spectrum Archive EE problem determination . . . . . . . . . . . . . . . . . . . . . . . . . . 261
7.24.1 Rsyslog log suppression by rate-limiting. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
7.25 Collecting IBM Spectrum Archive EE logs for support . . . . . . . . . . . . . . . . . . . . . . . 262
7.26 Backing up files within file systems that are managed by IBM Spectrum Archive EE 264
7.26.1 Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265
7.26.2 Backing up a GPFS or IBM Spectrum Scale environment . . . . . . . . . . . . . . . . 265
7.27 IBM TS4500 Automated Media Verification with IBM Spectrum Archive EE . . . . . . 266
7.28 How to disable commands on IBM Spectrum Archive EE. . . . . . . . . . . . . . . . . . . . . 271
7.29 LTO 9 Media Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 272
Contents vii
8.11 AFM use cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290
8.11.1 Centralized archive repository. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290
8.11.2 Asynchronous archive replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291
viii IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Notices
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
IBM® Interconnect® Redbooks®
IBM Cloud® Operating System/2® Redbooks (logo) ®
IBM Elastic Storage® POWER® Storwize®
IBM FlashSystem® PowerVM® Tivoli®
IBM Spectrum® ProtecTIER®
Intel, Intel Xeon, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks
of Intel Corporation or its subsidiaries in the United States and other countries.
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
Linear Tape-Open, LTO, Ultrium, the LTO Logo and the Ultrium logo are trademarks of HP, IBM Corp. and
Quantum in the U.S. and other countries.
Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.
Java, and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its
affiliates.
Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United States and
other countries.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Other company, product, or service names may be trademarks or service marks of others.
x IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Preface
This IBM® Redbooks® publication helps you with the planning, installation, and configuration
of the new IBM Spectrum® Archive Enterprise Edition (EE) Version 1.3.2.2 for the IBM
TS4500, IBM TS3500, IBM TS4300, and IBM TS3310 tape libraries.
IBM Spectrum Archive Enterprise Edition enables the use of the LTFS for the policy
management of tape as a storage tier in an IBM Spectrum Scale based environment. It also
helps encourage the use of tape as a critical tier in the storage environment.
This edition of this publication is the tenth edition of IBM Spectrum Archive Installation and
Configuration Guide.
IBM Spectrum Archive EE can run any application that is designed for disk files on a physical
tape media. IBM Spectrum Archive EE supports the IBM Linear Tape-Open (LTO) Ultrium 9,
8, 7, 6, and 5 tape drives. and the IBM TS1160, TS1155, TS1150, and TS1140 tape drives.
IBM Spectrum Archive EE can play a major role in reducing the cost of storage for data that
does not need the access performance of primary disk. The use of IBM Spectrum Archive EE
to replace disks with physical tape in tier 2 and tier 3 storage can improve data access over
other storage solutions because it improves efficiency and streamlines management for files
on tape. IBM Spectrum Archive EE simplifies the use of tape by making it transparent to the
user and manageable by the administrator under a single infrastructure.
This publication is intended for anyone who wants to understand more about IBM Spectrum
Archive EE planning and implementation. This book is suitable for IBM customers, IBM
Business Partners, IBM specialist sales representatives, and technical specialists.
Authors
This book was produced by a team working with the IBM Tokyo and IBM Tucson development
labs.
xii IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Larry Coyne is a Project Leader at the International Technical
Support Organization, Tucson Arizona center. He has over 35
years of IBM experience with 23 years in IBM storage software
management. He holds degrees in Software Engineering from
the University of Texas at El Paso and Project Management
from George Washington University. His areas of expertise
include customer relationship management, quality assurance,
development management, and support management for IBM
Storage Software.
Thanks to the following people for their contributions and support for this project:
Atsushi Abe
Hiroshi Itagaki
Takeshi Ishimoto
Osamu Matsumiya
Junta Watanabe
IBM Systems
Thanks to the authors of the previous editions of the IBM Spectrum Archive EE Redbooks:
Illarion Borisevich, Larry Coyne, Chris Hoffmann, Stefan Neff, Khanh Ngo, Wei Zheng Ong,
and Markus Schaefer
Find out more about the residency program, browse the residency index, and apply online at
this website:
http://www.ibm.com/redbooks/residencies.html
Preface xiii
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
http://www.ibm.com/redbooks
Send your comments in an email to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
xiv IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Summary of changes
This section describes the technical changes that are made in this edition of the book and in
previous editions. This edition might also include minor corrections and editorial changes that
are not identified.
Summary of Changes
for SG24-8333-09
for IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
as created or updated on March 10, 2022.
Note: Review the readme and important notice files in IBM Fix Central for the latest IBM
Spectrum Archive Enterprise Edition Version to make sure your upgrade plan will be
current with important recommended updates.
Updated the following sections in Chapter 3., “Planning for IBM Spectrum Archive Enterprise
Edition” on page 39.
Added deployment options to “IBM Spectrum Archive EE deployment options” on page 40
Updated “Data-access methods” on page 46
Reclaim tape drive usage has been updated in Chapter 6., “Managing daily operations of IBM
Spectrum Archive Enterprise Edition” on page 129 to describe how IBM Spectrum Archive
supports the parallel reclaim feature.
Added the following new use cases to Chapter 8., “IBM Spectrum Archive Enterprise Edition
use cases” on page 275.
8.4, “High-Performance Computing” on page 281
8.10, “S3 Object Interface” on page 288
xvi IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Enhanced the eeadm migrate, eeadm premigrate and eeadm save commands to accept a
list of files from stdin or from an input file.
Updated Red Hat Enterprise Linux Servers and IBM Spectrum Scale levels
Updated the required software
Added the description of ports used by IBM Spectrum Archive
Added the description of disabling/enabling the transparent recalls
Updated recommended inode size
Added limitation of file path length
Updated REST API definition
Renamed LE+ component to LE component
All examples and references have been updated with the new CLI syntax changes and
output. Added online help description for the eeadm CLI command.
Removed chapter about upgrading from 1.1.x
Changed Information
Updated Red Hat Enterprise Linux Servers and IBM Spectrum Scale levels
All examples and references have been updated with the new CLI syntax changes and
output
Changed Information
Updated Red Hat Enterprise Linux Servers and IBM Spectrum Scale levels
Updated various examples to include new information
Changed Information
Upgraded the HSM component
New information
Added support for Active File Management (AFM) Independent Writer (IW) mode starting
with IBM Spectrum Archive v1.2.3.0 and later
Added support for a RESTful API
Added high availability features:
– Control node failover
– Monitoring daemon
– New start/stop
Added a GUI dashboard for data monitoring
Added low pool threshold attribute for pools
Added support for 15 TB tape support with TS1155 tape drive
Added new ltfsee node show command
Added new ltfsee failover command
xviii IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Added new IBM Spectrum Archive EE database backup
Added IBM Swift HLM support
Changed Information
Added new traps to SNMP
Updated ltfsee info nodes command
Note: For more information about migration performance, see IBM Spectrum Archive
Enterprise Edition v1.2.2 Performance here.
Updated the following examples: Example 6-60 on page 169, Example 6-62 on page 170,
Example 6-63 on page 171, Example 6-64 on page 175, Example 6-66 on page 177,
Example 7-7 on page 244, Example 7-8 on page 245, Example 7-9 on page 246,
Example 7-19 on page 259.
New information
The following topics are new in Version 1.2.2:
Added new write failure state, Write Fenced
Added new pool remove option, -E for removing tapes with no file references
Increased stability and performance
Improved export/import
Improved reconcile
Automated the recover process of write failure tapes
Added improved method for recovering read failure tapes
Changed information
The following information was changed from the previous edition:
Added performance section
Renamed “Recovering data from a write-failure tape” to 6.16, “Reconciliation” on page 197
Updated the ltfsee recover command in 10.1, “Command-line reference” on page 316
Added section about recovering data from write failure tapes
Added section about recovering data from read failure tapes
Added section for handling Normal Export errors
New information
What’s New in Version 1.2.1:
Added procedures for upgrading from IBM Spectrum Archive Enterprise Edition version
1.1.x
Updated Red Hat Enterprise Linux Servers
Added information about avoiding errors during an upgrade on multiple nodes in 4.3.2,
“Installing, upgrading, or uninstalling IBM Spectrum Archive EE” on page 78
Changed information
Removed commands for creating a list of files to use with migration, premigration, and
save functions. v1.2.1.x and later now only uses the IBM GPFS scan result file.
Updated the options for the ltfsee_install command.
New information
What’s New in Version 1.2:
Multiple tape library attachment (up to two) support to a single IBM Spectrum Scale
cluster:
– Data replication to the pools in separate libraries for added data resiliency.
– You can use the read replica policy to specify file recall behavior if there is a failure.
– Total capacity expansion beyond a single library limit.
– IBM Spectrum Scale cluster in single site or across metro distance locations through
IBM Spectrum Scale synchronous mirroring.
– Mixed type library support with different drive types.
– Pools can be configured with a subset of tapes in one tape library as an LTO tape pool,
3592 tape pool, or 3592 write-once, read-many (WORM) tape pool.
xx IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Data recording on WORM tape cartridges:
– WORM cartridges are great for long-term records retention applications.
– Support for 3592 WORM cartridges (3592-JY and 3592-JZ tapes).
– Selection of rewritable or WORM cartridges per tape pool.
– Files can be migrated (or premigrated) to the specified WORM pool as the destination
in the IBM Spectrum Scale policy file or CLI option.
– Files on WORM tapes will not be overwritten and are not erasable:
• Reclamation, reconciliation, and reformat operations are disabled.
• As a preferred practice, set the immutable flag on IBM Spectrum Scale disk to
prevent the deletion of stub file for more data protection.
Expand storage capacity with LTO7 support:
– You have 2.4 - 4 times the capacity in the same foot print compared to LTO6 and LTO5
technology.
– Support for migration of files larger than 2.x TB (up to 6.x TB).
– Intermixing of LTO generation in single library is supported (the pool must be
homogeneous).
Performance improvement for large-scale systems:
– Optimization of file migration operations for many small files and for multi-node
configuration.
– Reduced IBM Spectrum Scale file system size requirements for IBM Spectrum Archive
EE metadata.
– Collocation of files within a pool for speedier file recall.
Flexibility in pool-based data management:
– Improved automatic recovery process on failure by switching to replica pools.
– Improved support of tape drive intermix.
New information
Added support for IBM GPFS V4.1
Added support for IBM TS1150 tape drive and media
v1.1.1.2 (PGA2.1) updates
Changed information
Added the -u option to the reconcile command to skip the pretest that checks for the
necessity to reconcile before mounting the tapes.
New information
Premigration commands
Added support for IBM TS4500 and TS3310 tape libraries
More operating system platform support
Preserving file system objects on tape
Recall commands
Version 1.1.1.0 (PGA1) and Version 1.1.1.1 (PGA2) updates
Changed information
Improved small file migration performance
Improved data resiliency (All copies are now referenced.)
xxii IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
1
This chapter includes the following topics for IBM Spectrum Archive Enterprise Edition (EE):
1.1, “Introduction” on page 2
1.2, “IBM Spectrum Archive EE functions” on page 6
1.3, “IBM Spectrum Archive EE components” on page 9
1.4, “IBM Spectrum Archive EE cluster configuration introduction” on page 15
IBM Spectrum Archive offers three software solutions for managing your digital files with the
LTFS format: Single Drive Edition (SDE), Library Edition (LE), and Enterprise Edition (EE).
This book focuses on the IBM Spectrum Archive EE.
IBM Spectrum Archive EE provides seamless integration of LTFS with IBM Spectrum Scale,
which is another member of the IBM Spectrum Storage family, by creating a tape-based
storage tier. You can run any application that is designed for disk files on tape by using IBM
Spectrum Archive EE because it is fully transparent and integrates in the IBM Spectrum Scale
file system. IBM Spectrum Archive EE can play a major role in reducing the cost of storage for
data that does not need the access performance of primary disk.
With IBM Spectrum Archive EE, you can enable the use of LTFS for the policy management
of tape as a storage tier in an IBM Spectrum Scale environment and use tape as a critical tier
in the storage environment.
The use of IBM Spectrum Archive EE to replace online disk storage with tape in tier 2 and tier
3 storage can improve data access over other storage solutions because it improves
efficiency and streamlines management for files on tape. IBM Spectrum Archive EE simplifies
the use of physical tape by making it not apparent to the user and manageable by the
administrator under a single infrastructure.
Figure 1-1 shows the integration of an IBM Spectrum Archive EE archive solution.
2 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
IBM Spectrum Archive EE uses the IBM Spectrum Archive LE for the movement of files to
and from tape devices. The scale-out architecture of IBM Spectrum Archive EE can add
nodes and tape devices as needed to satisfy bandwidth requirements between IBM Spectrum
Scale and the IBM Spectrum Archive EE tape tier.
Low-cost storage tier, data migration, and archive needs that are described in the following
use cases can benefit from IBM Spectrum Archive EE:
Operational storage
Provides a low-cost, scalable tape storage tier.
Active archive
A local or remote IBM Spectrum Archive EE node serves as a migration target for IBM
Spectrum Scale that transparently archives data to tape that is based on policies set by
the user.
The following IBM Spectrum Archive EE characteristics cover a broad base of integrated
storage management software with leading tape technology and the highly scalable IBM tape
libraries:
Integrates with IBM Spectrum Scale by supporting file-level migration and recall with an
innovative database-less storage of metadata.
Provides a scale-out architecture that supports multiple IBM Spectrum Archive EE nodes
that share tape inventory with load balancing over multiple tape drives and nodes.
Enables tape cartridge pooling and data exchange for IBM Spectrum Archive EE tape tier
management:
– Tape cartridge pooling allows the user to group data on sets of tape cartridges.
– Multiple copies of files can be written on different tape cartridge pools, including
different tape libraries in different locations.
– Supports tape cartridge export with and without the removal of file metadata from IBM
Spectrum Scale.
– Supports tape cartridge import with pre-population of file metadata in IBM Spectrum
Scale.
Tip: To learn more about IBM Spectrum Archive EE and try it in a virtual tape library
environment, see IBM Spectrum Archive Enterprise Edition Fundamentals & Lab Access.
Using an IBM Spectrum Archive tape tier as operational storage is useful when a significant
portion of files on a disk storage system infrastructure is static, meaning the data is not
changing.
In this case, as shown in Figure 1-2, it is optimal to move the content to a lower-cost storage
tier, in this case a physical tape. The files that are migrated to the IBM Spectrum Archive EE
tape tier remain online, meaning they are accessible from the IBM Spectrum Scale file system
under the IBM Spectrum Scale namespace at any time. Tape cartridge pools within IBM
Spectrum Archive EE can also be used for backup.
Gold
Silver
Bronze
Figure 1-2 Tiered operational storage with IBM Spectrum Archive EE managing the tape tier
4 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
With IBM Spectrum Archive EE, the user specifies files to be migrated to the IBM Spectrum
Archive tape tier by using standard IBM Spectrum Scale scan policies. IBM Spectrum Archive
EE then manages the movement of IBM Spectrum Scale file data to the IBM Spectrum
Archive tape cartridges. It also edits the metadata of the IBM Spectrum Scale files to point to
the content on the IBM Spectrum Archive tape tier.
Access to the migrated files through the IBM Spectrum Scale file system remains unchanged,
with the file data provided at the data rate and access times of the underlying tape
technology. The IBM Spectrum Scale namespace is unchanged after migration, making the
placement of files in the IBM Spectrum Archive tape tier not apparent to users and
applications. See 7.10.1, “Creating a traditional archive system policy” on page 244.
The use of an LTFS tape tier as an active archive is useful when you need a low-cost,
long-term archive for data that is maintained and accessed for reference. IBM Spectrum
Archive satisfies the needs of this type of archiving by using open-format, portable, and
self-describing tapes based on the LTFS standard.
In an active archive, the IBM Spectrum Archive file system is the main store for the data while
the IBM Spectrum Scale file system, with its limited disk capacity, is used as a staging area,
or cache, in front of IBM Spectrum Archive EE. IBM Spectrum Scale policies are used to
stage and de-stage data from the IBM Spectrum Scale disks to the IBM Spectrum Archive EE
tape cartridge.
Figure 1-3 shows the archive storage management with the IBM Spectrum Archive tape tier in
the IBM Spectrum Scale file system, the disk that is used for caching, and the namespace
that is mapped to the tape cartridge pool.
Disk cache
For more information see 7.10.2, “Creating active archive system policies” on page 245.
User data
IBM Spectrum
Archive LE
Figure 1-4 IBM Spectrum Archive EE integration with IBM Spectrum Scale and IBM Spectrum Archive LE
With IBM Spectrum Archive EE, you can perform the following management tasks on your
system:
Create and define tape cartridge pools for file migrations.
Migrate files in the IBM Spectrum Scale namespace to the IBM Spectrum Archive tape
tier.
Recall files that were migrated to the IBM Spectrum Archive tape tier back into IBM
Spectrum Scale.
Reconcile file inconsistencies between files in IBM Spectrum Scale and their equivalents
in IBM Spectrum Archive.
Reclaim tape space that is occupied by non-referenced files and non-referenced content
that is present on the physical tapes.
Export tape cartridges to remove them from your IBM Spectrum Archive EE system.
Import tape cartridges to add them to your IBM Spectrum Archive EE system.
Add tape cartridges to your IBM Spectrum Archive EE system to expand the tape
cartridge pool with no disruption to your system.
Obtain inventory, task status of your IBM Spectrum Archive EE solution.
6 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
1.2.1 User Task Reporting
Versions prior to IBM Spectrum Archive EE v1.3.0.0 reported any running or pending
operations through the ltfsee info scans and ltfsee info jobs commands. However any
successes or failures of prior commands, history of operations, or output of commands
cannot be seen except through examination of the trace logs. Furthermore, the result of any
file state changes from any migration, premigration, or recall commands cannot be viewed.
IBM Spectrum Archive EE v1.3.0.0 introduced the concept of user tasks. A user task is
defined as:
User initiated command or operation where each is considered a single task.
Each task is uniquely identified by a task ID in the range of 1,000 to 99,999 with the next
task ID after 99,999 starting back at 1,000.
A maximum number of tasks that can be accepted is defined as follows:
– 128 transparent recalls per each active control node
– For tasks other than transparent recall, 512 tasks for each active control node
If a task cannot be accepted due to this limit, the command is rejected immediately without
assigning a task ID.
Tasks can be manually cleared for any completed tasks.
The last 10 tasks with errors will always be preserved unless manually cleared.
8 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
– eeadm task clearhistory
– eeadm task cancel
Note: The other commands (eeadm migrate, eeadm premigrate, eeadm recall, and eeadm
save) does not fit well with the “--async” option when used through the mmapplypolicy
command because it will cause mmapplypolicy command to continuously submit the same
files repeatedly.
IBM Spectrum Scale is a required component for the IBM Spectrum Archive solution.
IBM Spectrum Archive EE is composed of multiple components that enable an IBM Spectrum
Archive tape tier to be used for migration and recall with the IBM Spectrum Scale. Files are
migrated to, and recalled from, the IBM Spectrum Archive tape tier by using the IBM
Spectrum Archive EE components that are shown in Figure 1-5 on page 10 and Figure 1-6 on
page 11.
Figure 1-5 shows the components that make up IBM Spectrum Archive EE. The components
are shown with IBM Spectrum Scale configured on separate nodes for maximum scalability.
IBM Spectrum Scale IBM Spectrum Scale IBM Spectrum Scale IBM Spectrum Scale
Tape Library
Figure 1-5 Components of IBM Spectrum Archive EE with separate IBM Spectrum Scale nodes
10 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
In Figure 1-6, the components that make up IBM Spectrum Archive EE are shown with no
separate IBM Spectrum Scale nodes. Also shown is how IBM Spectrum Scale can be
configured to run on the same nodes as the IBM Spectrum Archive EE nodes.
Tape Library
Figure 1-6 Components of IBM Spectrum Archive EE with no separate IBM Spectrum nodes
A second tape library can be added to the configuration, which expands the storage capacity
and offers the opportunity to add nodes and tape devices. Availability can be improved by
storing redundant copies on different tape libraries.
With multiple tape library attachments, the tape libraries can be connected to an IBM
Spectrum Scale cluster in a single site or can be placed in the metro distance locations
through IBM Spectrum Scale synchronous mirroring (stretched cluster).
For IBM Spectrum Scale V5.0.0 and later, distances up to 300 km (186.4 miles) are
supported for stretched cluster with synchronous mirroring using block-level replication.
It is important to remember that this is still a single IBM Spectrum Scale cluster. In a
configuration using IBM Spectrum Scale replication, a single IBM Spectrum Scale cluster is
defined over two geographically separated sites consisting of two active production sites and
by using tiebreaker disks.
One or more file systems are created, mounted, and accessed concurrently from the two
active production sites. The data and metadata replication features of IBM Spectrum Scale
are used to maintain a secondary copy of each file system block, relying on the concept of
disk failure groups to control the physical placement of the individual copies:
1. Separate the set of available disk volumes into two failure groups. Define one failure group
at each of the active production sites.
2. Create a replicated file system. Specify a replication factor of 2 for both data and
metadata.
For more information about synchronous mirroring that uses IBM Spectrum Scale replication,
see IBM Documentation.
Important: Stretched cluster is available for distances of 100 km (62 miles) - 300 km
(186.4 miles). For longer distances, use the AFM feature of IBM Spectrum Scale with IBM
Spectrum Archive.
With the release of IBM Spectrum Archive v1.2.3.0, limited support is provided for IBM
Spectrum Scale AFM. The only supported AFM is with two different IBM Spectrum Scale
clusters with one instance of IBM Spectrum Archive at each site.
For more information about IBM Spectrum Scale AFM, see 2.2.5, “Active File
Management” on page 30.
GPFSNative
GPFS Native Client GPFSNative
NativeClient
Client GPFS Native Client GPFSNative
Native Client
NFS ClientClient GPFS
CIFS/SMB Client IBM GPFS Native
Spectrum Client
Scale Client GPFS
FTP ClientClient
Ethernet
F F F F F F F F F F F F
SAN or
shared NSD access
D D D D D D D D D D D D
Library 1 Library 2
Pool 1 primary Pool 1 secondary
Legend
Network Connection
IBM Spectrum Archive Node F Fibre Channel or SAS Port Tape Cartridge
SAN Connection
Tape Library (logical) D Tape Drive Disk or Flash Storage Tape Drive Connection
12 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
1.3.2 Hierarchical Storage Manager
A Hierarchical Storage Manager (HSM) solution typically moves the file’s data to back-end
storage (in most cases, physical tape media) and leaves a small stub file in the local storage
file system. The stub file uses minimal space, but leaves all metadata information about the
local storage in such a way that for a user or a program the file looks like a normal, local
stored file. When the user or a program accesses the file, the HSM solution automatically
recalls (moves back) the file’s data from the back-end storage and gives the reading
application access to the file when all the data is back retrieved and available online again.
The scheduler component of the control node uses policy-based cartridge selection to
schedule and process task requests, such as migration and recall requests, which are fulfilled
by using available system nodes and tape resources. The following tasks are performed by
the scheduler:
Choosing a task from the task queue
Choosing an appropriate tape cartridge and tape drive to handle the work
Starting the task
The control node also manages the creation of replicas across multiple tape libraries. For
example, when the eeadm migrate command specifies making replicas of files in multiple tape
libraries, the command accesses the control node that manages the tape pool for the primary
copy.
The control node puts the copy job in the task queue for the primary tape library, then passes
the secondary copy task to the control node for the second tape library. The second control
node puts the copy task in the task queue for the second tape library.
When the scheduler component of the control node selects a tape cartridge and tape drive for
a migration task, it manages the following conditions:
If the migration is to a tape cartridge pool, the tape drive must belong to the node group
that owns the tape cartridge pool.
If a format generation property is defined for a tape cartridge pool, the tape cartridge must
be formatted as that generation, and the tape drive must support that format.
The number of tape drives that are being used for migration to a tape cartridge pool at one
time must not exceed the defined mount limit.
If there are multiple candidate tapes available for selection, the scheduler tries to choose a
tape cartridge that is already mounted on an available tape drive.
When the scheduler selects a tape cartridge and tape drive for tasks other than migration, it
makes the following choices:
Choosing an available tape drive in the node group that owns the tape cartridge and tape
cartridge pool
Choosing the tape drive that has the tape drive attribute for the task
Other functions that are provided by the control node include the following functions:
Maintains a catalog of all known drives that are assigned to each IBM Spectrum Archive
node in the system
Maintains a catalog of tape cartridges in the tape library/libraries
Maintains an estimate of the free space on each tape cartridge
Allocates space on tape cartridges for new data
The MMM service is started when IBM Spectrum Archive EE is started by running the
eeadm cluster start command. The MMM service runs on only one IBM Spectrum Archive
EE control node, for each library, at a time. Several operations, including migration and recall,
fail if the MMM service stops. If SNMP traps are enabled, a notification is sent when the MMM
service starts or stops.
For more information, see 6.4, “Starting and stopping IBM Spectrum Archive EE” on
page 146, and 6.24, “Monitoring the system with SNMP” on page 216.
Important: If the eeadm cluster start command does not return after several minutes, it
might be because the firewall is running. The firewall service must be disabled on the IBM
Spectrum Archive EE nodes. For more information, see 4.3.2, “Installing, upgrading, or
uninstalling IBM Spectrum Archive EE” on page 78.
The eeadm cluster start command also does the unmount of the tape drives, so the
process might take a long time if there are many mounted tape drives.
The LE component is installed on all of the IBM Spectrum Scale nodes that are connected to
the IBM Spectrum Archive EE library. It is the migration target for IBM Spectrum Scale. The
LE component accesses the recording space on the physical tape cartridges through its file
system interface and handles the user data as file objects and associated metadata in its
namespace.
With IBM Spectrum Archive EE v1.3.0.0 and later, IBM Spectrum Archive LE is started
automatically when running the eeadm cluster start command. If errors occur during start of
the IBM Spectrum Archive EE system, run the eeadm node list command to display which
component failed to start.
For more information about the updated eeadm node list command, see 6.7, “IBM Spectrum
Archive EE automatic node failover” on page 153.
14 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
1.4 IBM Spectrum Archive EE cluster configuration
introduction
This section describes a cluster configuration for IBM Spectrum Archive EE. This
configuration is for single-library, multiple-node access.
Single-library, multiple-node access enables access to the same set of IBM Spectrum Archive
EE tape cartridges from more than one IBM Spectrum Archive EE node. The purpose of
enabling this capability is to improve data storage and retrieval performance by assigning
fewer tape drives to each node.
When this cluster configuration is used, each IBM Spectrum Archive EE node must have its
own set of drives that is not shared with any other node. In addition, each IBM Spectrum
Archive EE node must have at least one control path drive that is designated as a control path
by an operator of the attached IBM tape library.
IBM Spectrum Archive EE uses the drive that is designated as a control path to communicate
with the tape library. This type of control path is also known as a media changer device. IBM
Spectrum Archive EE is scalable so you can start out with a single node and add nodes later.
Important: As part of your planning, work with your IBM tape library administrator to
ensure that each IBM Spectrum Archive EE node in your configuration has its own media
changer device (control path) defined in its logical library.
SAN or
shared NSD access
Tape Library
Robot
Tape
IBM Spectrum Archive EE manages all aspects of the single-library, multiple-node access,
which includes the management of the following areas:
Multiple tenancy
The contents of the tape cartridge are managed automatically by the IBM Spectrum
Archive EE system so that each IBM Spectrum Archive EE node does not have to be
aware of any changes made on other IBM Spectrum Archive EE nodes. The index on each
tape cartridge is updated when the tape is mounted and the index is read from this tape.
Single node management of library inventory
The IBM Spectrum Archive EE system automatically keeps the library inventory up to date
to manage the available drives and tape cartridges. The library inventory is kept on the
node on which the MMM service runs.
Space reclaim management
When data is moved from one tape cartridge to another to reclaim the space on the first
tape cartridge, the IBM Spectrum Archive EE system ensures that the physical changes
on the cartridges are handled correctly.
16 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2
Note: Throughout this publication, the terms supported tape libraries, supported tape
drives, and supported tape media are used to represent the following tape libraries, tape
drives, and tape media. Unless otherwise noted, as of the date of publication, IBM
Spectrum Archive EE supports these libraries:
IBM TS4500 tape librarya
IBM TS4300 tape library
IBM TS3500 tape libraryab
IBM TS3310 tape library
IBM LTO Ultrium 9, 8, 7, 6, or 5 tape drives, and IBM TS1160 (3592 60E,60F, 60G, or
60S), TS1155 (3592 55F or 55G), TS1150, or TS1140 tape drives
LTO Ultrium 9, 8, M8c, 7, 6, and 5, and 3592 JB, JC, JD, JE, JK, JL, JM, JV, JY, and JZ
tape media
For more information about the latest system requirements, see IBM Documentation.
For the latest IBM Spectrum Archive Library Edition Support Matrix (supported tape library
and tape drive firmware levels, see IBM Support’s Fix Central web page.
a. IBM TS1160, TS1155, TS1150, and IBM TS1140 support on IBM TS4500 and IBM TS3500
tape libraries only.
b. TS3500 does not support the LTO-9 tape drive.
c. Uninitialized M8 media (MTM 3589-452) is only supported on the following tape libraries in IBM
Spectrum Archive EE: TS4500, TS4300, and TS3310. TS3500 will only support pre-initialized
media.
With this application, accessing data that is stored on an IBM tape cartridge is as easy and
intuitive as the use of a USB flash drive. Tapes are self-describing, and you can quickly recall
any file from a tape cartridge without having to read the whole tape cartridge from beginning
to end. Furthermore, any LTFS-capable system can read a tape cartridge that is created by
any other LTFS-capable system (regardless of the operating system). Any LTFS-capable
system can identify and retrieve the files that are stored on it. LTFS-capable systems have the
following characteristics:
Files and directories are shown to you as a directory tree listing.
More intuitive searches of tape cartridges and library content are now possible because of
the addition of file tagging.
Files can be moved to and from LTFS tape cartridges by using the familiar drag method
that is common to many operating systems.
Many applications that were written to use files on disk can now use files on tape
cartridges without any modification.
All standard File Open, Write, Read, Append, Delete, and Close functions are supported.
No need for another external tape management system or database tracking the content
of each tape.
18 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Archival data storage requirements are growing at over 60% annually. The LTFS format is an
ideal option for long-term archiving of large files that must be easily shared with others. This
option is especially important because the tape media that it uses (LTO and 3592) are
designed to have a 10+ years lifespan (depending on the number of read/write passes).
Industries that most benefit from this tape file system are the banking, digital media, medical,
geophysical, and entertainment industries. Many users in these industries use Linux or
Macintosh systems, which are fully compatible with LTFS.
LTO Ultrium tape cartridges from earlier LTO generations (that is, LTO-1 through LTO-4)
cannot be partitioned and be used by LTFS/IBM Spectrum Archive. Also, if LTO Ultrium 4 tape
cartridges are used in an LTO Ultrium 5 tape drive to write data, the LTO-4 tape cartridge is
treated as an unpartitioned LTO-5 tape cartridge. Even if an application can manage
partitions, it is not possible to partition the LTO-4 media that is mounted in an LTO Ultrium 5
drive.
Starting with the release of IBM Spectrum Archive EE v1.2, corresponding Write Once, Read
Many (WORM) tape cartridges are supported in an IBM Spectrum Archive EE solution that
operates supported IBM Enterprise tape drives. With the same release, tape drives in mixed
configurations are supported. For more information, see 6.21, “Tape drive intermix support”
on page 208.
Although LTFS presents the tape cartridge as a disk drive, the underlying hardware is still a
tape cartridge and is therefore sequential in nature. Tape does not allow random access. Data
is always appended to the tape, and there is no overwriting of files. File deletions do not erase
the data from tape, but instead erase the pointers to the data. So, although with LTFS you can
simultaneously copy two (or more) files to an LTFS tape cartridge, you get better performance
if you copy files sequentially.
To operate the tape file system, the following components are needed:
Software in the form of an open source LTFS package
Data structures that are created by LTFS on tape
Together, these components can manage a file system on the tape media as though it is a
disk file system for accessing tape files, including the tape directory tree structures. The
metadata of each tape cartridge, after it is mounted, is cached to the server. Therefore,
metadata operations, such as browsing the directory or searching for a file name, do not
require any tape movement and are quick.
IBM TS1160 tape Advanced type E data (JE) 20000 GB (18626 GiB) 19485 GB (18147 GiB)
drivee, f
Advanced type D data (JD) 15000 GB (13969 GiB) 14562 GB (13562 GiB)
Advanced Type C data (JC) 7000 GB (6519 GiB) 6757 GB (6293 GiB)
IBM TS1155 tape Advanced type D data (JD) 15000 GB (13969 GiB) 14562 GB (13562 GiB)
drivee, f
Advanced type C data (JC) 7000 GB (6519 GiB) 6757 GB (6293 GiB)
IBM TS1150 tape Advanced type D data (JD) 10,000 GB (9313 GiB) 9687 GB (9022 GiB)
drivee,f
Advanced type C data (JC) 7000 GB (6519 GiB) 6757 GB (6293 GiB)
20 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Tape drive Tape mediaa Native capacityb, c LTFS data partition sizecd
IBM TS1140 tape Advanced data type C (JC) 4000 GB (3725 GiB) 3650 GB (3399 GiB)
drivee, f
Advanced data type C 4000 GB (3725 GiB) 3650 GB (3399 GiB)
WORM (JY)
IBM LTO Ultrium 9 LTO 9 18000 GB (16764 GiB) 17549 GB (16344 GiB)
tape drive
IBM LTO Ultrium 8 LTO 8 12000 GB (11175 GiB) 11711 GB (10907 GiB)
tape drive
LTO 8 (M8) 9000 GB (8382 Gib) 8731 GB (8132 GiB)
IBM LTO Ultrium 7 LTO 7 6000 GB (5588 GiB) 5731 GB (5338 GiB)
tape drive
IBM LTO Ultrium 6 LTO 6 2500 GB (2328 GiB) 2408 GB (2242 GiB)
tape drive
IBM LTO Ultrium 5 LTO 5 1500 GB (1396 GiB) 1425 GB (1327 GiB)
tape drive
a. WORM media are not supported by IBM Spectrum Archive SDE and IBM Spectrum Archive LE, only with EE.
b. The actual usable capacity is greater when compression is used.
c. See the topic Data storage values, found here.
d. Values that are given are the default size of the LTFS data partition, unless otherwise indicated.
e. TS1160, TS1155, TS1150, and TS1140 tape drives support enhanced partitioning for cartridges.
f. Media that are formatted on a 3592 drive must be read on the same generation of drive. For example, a JC
cartridge that was formatted by a TS1150 tape drive cannot be read on a TS1140 tape drive.
Note: IBM LTFS Storage Manager (LTFS SM) was discontinued from marketing effective
12/14/2015. IBM support for the LTFS SM was discontinued 05/01/2020.
Figure 2-1 shows the IBM Spectrum Archive SDE user view, which resembles standard file
folders.
It runs on Linux, Windows, and MacOS, and with the operating system’s graphical File
Manager, reading data on a tape cartridge is as easy as dragging file data sets. Users can
run any application that is designed for disk files against tape data without concern for the fact
that the data is physically stored on tape. IBM Spectrum Archive SDE allows access to all of
the data in a tape cartridge that is loaded on a single drive as though it were on an attached
disk drive.
It supports stand-alone versions of LTFS, such as those running on IBM, HP, Quantum,
FOR-A, 1 Beyond, and other platforms.
IBM Spectrum Archive SDE software, systems, tape drives and media
requirements
The most current software, systems, tape drives and media requirements can be found at the
IBM Spectrum Archive Single Drive Edition IBM Documentation web page.
Select the most current IBM Spectrum Archive SDE version and then select Planning. The
Supported tape drives and media, system requirements, and required software topics are
displayed.
IBM Spectrum Archive SDE supports the use of multiple tape drives at one time. The method
for using multiple tape drives depends on the operating system being used.
22 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
For Linux and Mac OS X users, it is possible to use multiple tape drives by starting multiple
instances of the LTFS software, each with a different target tape device name in the -o
devname parameter. For more information, see the Mounting media by using the ltfs command
topic at IBM Documentation.
For Windows users, the LTFS software detects each of the installed tape drives, and it is
possible to assign a different drive letter to each drive by using the configuration window. For
more information, see the Assigning a drive letter to a tape drive topic at IBM Documentation.
Note: A certain level of tape drive firmware is required to fully use IBM Spectrum Archive
SDE functions. To find the supported firmware version and for more information about
connectivity and configurations, see the IBM System Storage Interoperation Center (SSIC)
web page.
Data tapes that are used by IBM Spectrum Archive SDE version 1.3.0 or later can be
imported into IBM Spectrum Archive EE. For more information about this procedure, see
6.19.1, “Importing” on page 202. Tapes that were formatted in LTFS 1.0 format by older
versions of IBM Spectrum Archive are automatically upgraded to LTFS 2.4 format on first
write.
Figure 2-2 shows the user view of multiple IBM Spectrum Archive tapes appearing as
different library folders.
Figure 2-2 IBM Spectrum Archive LE User view of multiple LTFS tape cartridges
The LTFS metadata of each tape cartridge, after it is mounted, is cached in server memory.
So, even after the tape cartridge is ejected, the tape cartridge metadata information remains
viewable and searchable, with no remounting required. Every tape cartridge and file is
accessible through the operating system file system commands, from any application. This
improvement in search efficiency can be substantial, considering the need to search
hundreds or thousands of tape cartridges that are typically found in tape libraries.
Select the most current IBM Spectrum Archive LE version and then select Planning. The
Supported tape drives and media, system requirements, and required software topics will be
displayed.
For more information about connectivity and configurations, see the SSIC website.
Data tapes that were created and used by IBM Spectrum Archive LE Version 2.1.2 or later
can be imported into IBM Spectrum Archive EE. For more information about this procedure,
see 6.19.1, “Importing” on page 202.
IBM Spectrum Archive EE provides seamless integration of LTFS with IBM Spectrum Scale
by providing an IBM Spectrum Archive tape tier under IBM Spectrum Scale. IBM Spectrum
Scale policies are used to move files between online disks storage and IBM Spectrum Archive
tape tiers without affecting the IBM Spectrum Scale namespace.
IBM Spectrum Archive EE uses IBM Spectrum Archive LE for the movement of files to and
from the physical tape devices and cartridges. IBM Spectrum Archive EE can manage
multiple IBM Spectrum Archive LE nodes in parallel, so bandwidth requirements between IBM
Spectrum Scale and the tape tier can be satisfied by adding nodes and tape devices as
needed.
24 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 2-3 shows the IBM Spectrum Archive EE system view with IBM Spectrum Scale
providing the global namespace and IBM Spectrum Archive EE installed on two IBM
Spectrum Scale nodes. IBM Spectrum Archive EE can be installed on one or more IBM
Spectrum Scale nodes. Each IBM Spectrum Archive EE instance has dedicated tape drives
that are attached in the same tape library partition. IBM Spectrum Archive EE instances
share tape cartridges and LTFS index. The workload is distributed over all IBM Spectrum
Archive EE nodes and their attached tape drives.
Global Namespace
IBM Spectrum Scale Node 1 IBM Spectrum Scale Node 2
A local or remote IBM Spectrum Archive LE node serves as a migration target for IBM
Spectrum Scale, which transparently archives data to tape based on policies set by the user.
Figure 2-4 provides a conceptual overview of processes and data flow in IBM Spectrum
Archive EE.
IBM Spectrum Archive EE can be used for a low-cost storage tier, data migration, and archive
needs as described in the following use cases.
Operational storage
The use of an IBM Spectrum Archive tape tier as operational storage is useful when a
significant portion of files on an online disk storage system is static, meaning the data does
not change. In this case, it is more efficient to move the content to a lower-cost storage tier, for
example, to a physical tape cartridge. The files that are migrated to the IBM Spectrum Archive
tape tier remain online, meaning they are accessible at any time from IBM Spectrum Scale
under the IBM Spectrum Scale namespace.
With IBM Spectrum Archive EE, the user specifies files to be migrated to the IBM Spectrum
Archive tape tier by using standard IBM Spectrum Scale scan policies. IBM Spectrum
Archive EE then manages the movement of IBM Spectrum Scale file data to IBM Spectrum
Archive tape cartridges. It also edits the metadata of the IBM Spectrum Scale files to point to
the content on the IBM Spectrum Archive tape tier.
Access to the migrated files through the IBM Spectrum Scale file system remains unchanged
with the file data provided at the data rate and access times of the underlying tape
technology. The IBM Spectrum Scale namespace is unchanged after migration, which makes
the placement of files in the IBM Spectrum Archive tape tier not apparent to users and
applications.
26 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Active archive
The use of an IBM Spectrum Archive tape tier as an active archive is useful when there is a
need for a low-cost, long-term archive for data that is maintained and accessed for reference.
IBM Spectrum Archive satisfies the needs of this type of archiving by using open-format,
portable, self-describing tapes. In an active archive, the LTFS-file system is the main storage
for the data. The IBM Spectrum Scale file system, with its limited disk capacity, is used as a
staging area, or cache, in front of IBM Spectrum Archive.
IBM Spectrum Scale policies are used to stage and destage data from the IBM Spectrum
Scale disks space to the IBM Spectrum Archive tape cartridges. The tape cartridges from the
archive can be exported for vaulting or for moving data to another location. Because the
exported data is in the LTFS format, it can be read on any LTFS-compatible system.
The entire file system is striped across all storage devices, typically disk and flash storage
subsystems.
Note: IBM Spectrum Scale is a member of the IBM Spectrum product family. The most
recent version at the time of writing is Version 5.1.2. The first version with the name IBM
Spectrum Scale is Version 4.1.1. The prior versions, Versions 4.1.0 and 3.5.x, are still
called GPFS. Both IBM Spectrum Scale and GPFS are used interchangeably in this book.
For more information about current documentation and publications for IBM Spectrum Scale,
see this IBM Documentation web page.
2.2.1 Overview
IBM Spectrum Scale can help you achieve Information Lifecycle Management (ILM)
efficiencies through powerful policy-driven automated tiered storage management. The IBM
Spectrum Scale ILM toolkit helps you manage sets of files, pools of storage, and automate
the management of file data. By using these tools, IBM Spectrum Scale can automatically
determine where to physically store your data regardless of its placement in the logical
directory structure. Storage pools, file sets, and user-defined policies can match the cost of
your storage resources to the value of your data.
You can use IBM Spectrum Scale policy-based ILM tools to perform the following tasks:
Create storage pools to provide a way to partition a file system’s storage into collections of
disks or a redundant array of independent disks (RAID) with similar properties that are
managed together as a group.
IBM Spectrum Scale has the following types of storage pools:
– A required system storage pool that you create and manage through IBM Spectrum
Scale.
– Optional user storage pools that you create and manage through IBM Spectrum Scale.
There are two types of storage pools in an IBM Spectrum Scale environment: Internal storage
pools and external storage pools. Internal storage pools are managed within IBM Spectrum
Scale. External storage pools are managed by an external application, such as IBM Spectrum
Archive EE. For external storage pools, IBM Spectrum Scale provides tools that you can use
to define an interface that IBM Spectrum Archive EE uses to access your data.
IBM Spectrum Scale does not manage the data that is placed in external storage pools.
Instead, it manages the movement of data to and from external storage pools. You can use
storage pools to perform complex operations such as moving, mirroring, or deleting files
across multiple storage devices, which provide storage virtualization and a single
management context.
Internal IBM Spectrum Scale storage pools are meant for managing online storage resources.
External storage pools are intended for use as near-line storage and for archival and backup
operations. However, both types of storage pools provide you with a method to partition file
system storage for the following considerations:
Improved price-performance by matching the cost of storage to the value of the data
Improved performance by:
– Reducing the contention for premium storage
– Reducing the impact of slower devices
– Allowing you to retrieve archived data when needed
Improved reliability by providing for:
– Replication based on need
– Better failure containment
– Creation of storage pools as needed
28 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2.2.3 Policies and policy rules
IBM Spectrum Scale provides a means to automate the management of files by using policies
and rules. If you correctly manage your files, you can use and balance efficiently your
premium and less expensive storage resources. IBM Spectrum Scale supports the following
policies:
File placement policies are used to place automatically newly created files in a specific
storage pool.
File management policies are used to manage files during their lifecycle by moving them
to another storage pool, moving them to near-line storage, copying them to archival
storage, changing their replication status, or deleting them.
A policy is a set of rules that describes the lifecycle of user data that is based on the file’s
attributes. Each rule defines an operation or definition, such as migrate to a pool and replicate
the file. The rules are applied for the following uses:
Initial file placement
File management
Restoring file data
When a file is created or restored, the placement policy determines the location of the file’s
data and assigns the file to a storage pool. All data that is written to that file is placed in the
assigned storage pool. The placement policy that is defining the initial placement of newly
created files and the rules for placement of restored data must be installed into IBM Spectrum
Scale by running the mmchpolicy command. If an IBM Spectrum Scale file system does not
have a placement policy that is installed, all the data is stored into the system storage pool.
Only one placement policy can be installed at a time.
If you switch from one placement policy to another, or change a placement policy, that action
has no effect on files. However, newly created files are always placed according to the
currently installed placement policy.
The management policy determines file management operations, such as migration and
deletion. To migrate or delete data, you must run the mmapplypolicy command. You can
define the file management rules and install them in the file system together with the
placement rules. As an alternative, you can define these rules in a separate file and explicitly
provide them to mmapplypolicy by using the -P option. In either case, policy rules for
placement or migration can be intermixed. Over the life of the file, data can be migrated to a
different storage pool any number of times, and files can be deleted or restored.
With Version 3.1, IBM Spectrum Scale introduced the policy-based data management that
automates the management of storage resources and the data that is stored on those
resources. Policy-based data management is based on the storage pool concept. A storage
pool is a collection of disks or RAIDs with similar properties that are managed together as a
group. The group under which the storage pools are managed together is the file system.
IBM Spectrum Scale provides a single name space across all pools. Files in the same
directory can be in different pools. Files are placed in storage pools at creation time by using
placement policies. Files can be moved between pools based on migration policies and files
can be removed based on specific policies.
For more information about the SQL-like policy rule language, see IBM Spectrum Scale:
Administration Guide, which is available at this IBM Documentation web page.
For more information about GPFS policies, see 6.11, “Migration” on page 165.
There are two different approaches that can be used to drive an IBM Spectrum Archive EE
migration through GPFS policies: Manual and automated. These approaches are only
different in how the mmapplypolicy command (which performs the policy scan) is started.
Manual
The manual IBM Spectrum Scale driven migration is performed when the user or a UNIX cron
job runs the mmapplypolicy command with a predefined migration or premigration policy. The
rule covers the migration or premigration of files from the system pool to the external IBM
Spectrum Scale pool, which means that the data is physically moved to the external tape
pool, which must be defined in IBM Spectrum Archive EE.
Automated
The GPFS threshold migration is performed when the user specifies a threshold policy and
the GPFS policy daemon is enabled to monitor the storage pools in the file system for that
threshold. If a predefined high threshold is reached (which means the filling level of the
storage pool reached the predefined high water mark), the monitor daemon automatically
starts the mmapplypolicy command to perform an inode scan.
For more information about migration, see 6.11, “Migration” on page 165.
Another way to get files transferred from home to cache is through prefetching. Prefetching
can use the IBM Spectrum Scale policy engine to quickly identify files that match certain
criteria.
When files are created or changed in cache, they can be replicated back to home. A file that
was replicated back to home can be evicted in cache. In this case, the user still sees the file in
cache (the file is uncached), but the actual file content is stored in home. Eviction is triggered
by the quota that is set on the AFM file set and can evict files based on size or last recent
used criteria.
30 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Cache must be an IBM Spectrum Scale independent file set. Home can be an IBM Spectrum
Scale file system, a Network File System (NFS) export from any other file system, or a file
server (except for the disaster-recovery use case). The caching relationship between home
and cache can be based on the NFS or native IBM Spectrum Scale protocol. In the latter
case, home must be an IBM Spectrum Scale file system in a different IBM Spectrum Scale
cluster. The examples in this Redpaper that feature AFM use an NFS protocol at the home
cluster.
The AFM relation is typically configured on the cache file set in one specific mode. The AFM
mode determines where files can be processed (created, updated, and deleted) and how files
are managed by AFM according to the file state. See Figure 2-5 on page 31 for AFM file
states.
Important: IBM Spectrum Archive EE V1.2.3.0 is the first release that started supporting
IBM Spectrum Scale AFM with IBM Spectrum Scale V4.2.2.3. AFM has multiple cache
modes that can be created. However, IBM Spectrum Archive EE supports only the
independent-writer (IW) cache mode.
Independent-writer
The IW cache mode of AFM makes the AFM target home for one or more caches. All
changes in the caches are replicated to home asynchronously. Changes to the same data are
applied in the home file set so that the changes are replicated from the caches. There is no
cross-cache locking. Potential conflicts must be resolved at the respective cache site.
A file in the AFM cache can have different states as shown in Figure 2-5. File states can be
different depending on the AFM modes.
Uncached
(Not with AFM Cached Dirty
DR mode)
Uncached
When an AFM relation is created between cache and home, and files are available in home,
these files can be seen in cache without being present. This state means that the file
metadata is present in the cache, but the file content is still on home. Such files are in status
uncached. In addition, the uncached status is achieved by evicting files from the cache.
Cached
When an uncached file is accessed in cache for a read or write operation, the file is fetched
from home. Fetching is the process of copying a file from home to cache. Files fetched from
home to cache are in cached state. Another way to fetch files from home to cache is by using
the AFM prefetch command (mmafmctl prefetch). This command can use the policy engine
to identify files quickly, according to certain criteria.
For information about how to configure AFM with IBM Spectrum Archive EE, see 7.10.3, “IBM
Spectrum Archive EE migration policy with AFM” on page 246. For use cases, see Figure
8-11, “IBM Spectrum Archive use case for university Scientific Data Archive” on page 286.
The idea of SOBAR is to have all the file data pre-migrated or migrated to the IBM Spectrum
Protect server and periodically back up the file system configuration and metadata of all the
directories and files into the IBM Spectrum Protect server.
When the file system recovery is needed, the file system is restored by using the
configuration, and all the directories or files are restored by using the metadata. After the
recovery, the files are in stub format (HSM migrated state) and can be recalled from the IBM
Spectrum Protect server on-demand.
Because the recovery processes only the metadata and does not involve recalling or copying
the file data, the recovery is fast with significantly reduced recovery time objective (RTO). This
is especially true for large file systems.
As of this writing, SOBAR is not supported on the IBM Spectrum Archive managed file
systems for DR. However, from version 1.3.1.2, IBM Spectrum Archive supports a procedure
that uses SOBAR to performed a planned data migration between two file systems.
By using SOBAR, a file system can be migrated without recalling the file data from tapes. For
more information, see 6.27, “File system migration” on page 231.
32 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2.3 OpenStack SwiftHLM
The Swift High Latency Media (SwiftHLM) project seeks to create a high-latency storage back
end that makes it easier for users to perform bulk operations of data tiering within a Swift data
ring. SwiftHLM enables IBM Spectrum Scale, IBM Spectrum Archive, and IBM Spectrum
Protect as the key products for this software-defined hybrid storage with object interface to
tape technology. Data is produced at significantly higher rates than a decade ago.
The storage and data management solutions of the past can no longer keep up with the data
demands of today. The policies and structures that decide and execute how that data is used,
discarded, or retained determines how efficiently the data is used. The need for intelligent
data management and storage is more critical now than ever before.
The Swift HLM project addresses this challenge. Running OpenStack Swift on top of HLM
storage allows you to cheaply store and efficiently access large amounts of infrequently used
object data. Data that is stored on tape storage can be easily adopted to an Object Storage
data interface. SwiftHLM can be added to OpenStack Swift (without modifying Swift) to
extend Swift’s interface.
This ability allows users to explicitly control and query the state (on disk or on HLM) of Swift
object data, including efficient pre-fetch of bulk objects from HLM to disk when those objects
must be accessed. This function, previously missing in Swift, provides similar functions as
Amazon Glacier does through the Glacier API or the Amazon S3 Lifecycle Management API.
BDT Tape Library Connector (open source) and IBM Spectrum Archive or IBM Spectrum
Protect are examples of HLM back ends that provide important and complex functions to
manage HLM resources (tape mounts and unmounts to drives, serialization of requests for
tape media, and tape drive resources). They can use SwiftHLM functions for a proper
integration with Swift.
Although access to data that is stored on HLM can be done transparently without the use of
SwiftHLM, this process does not work well in practice for many important use cases and other
reasons. SwiftHLM function can be orthogonal and complementary to Swift (ring to ring)
tiering (source). The high-level architecture of the low cost, high-latency media storage
solution is shown in Figure 2-6 on page 34.
For more information, see Implementing OpenStack SwiftHLM with IBM Spectrum Archive EE
or IBM Spectrum Protect for Space Management, REDP-5430.
OpenStack Swift
migrate
Primary Storage Archival Storage
High Latency
HDD
Media
recall
This monitoring feature consists of multiple components that are installed in the IBM
Spectrum Archive EE nodes as well as dedicating an external node for displaying the
dashboard.
34 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Logstash is used for data collection, and should be installed on all IBM Spectrum Archive EE
nodes. The data that is collected by Logstash is then sent to Elasticsearch on the external
monitoring node where it can query data quickly and send it to the Grafana component for
visualization. Figure 2-7 shows the IBM Spectrum Archive EE Dashboard architecture.
The Dashboard views are System Health, Storage, Activity, Config, and Task. Figure 2-8
shows an example of the IBM Spectrum Archive EE Dashboard Activity view.
For more information on configuring the Dashboard within your environment, see the IBM
Spectrum Archive Enterprise Edition Dashboard Deployment Guide, which is available at this
IBM Documentation web page.
They are equivalent to what the eeadm task list and eeadm task show commands display.
With the REST API, you can automate these queries and integrate the information into your
applications including the web/cloud.
For installation instructions, see 4.4, “Installing a RESTful server” on page 83.
For usage examples including commonly used parameters, see 6.26, “IBM Spectrum Archive
REST API” on page 219.
Archiving with no file deletion is the typical process that is used by many backup and archive
software products. In the case of IBM Spectrum Protect, an archive creates a copy of one or
more files in IBM Spectrum Protect with a set retention period. It is often used to create a
point-in-time copy of the state of a server’s file system and this copy is kept for an extended
period. After the archive finishes, the files are still on the server’s file system.
Contrast this with archiving with file deletion where after the archive finishes the files that form
part of the archive are deleted from the file system. This is a feature that is offered by the
IBM Spectrum Protect archive process. Rather than a point-in-time copy, it can be thought of
as a point-in-time move as the files are moved from the servers’ file system into IBM
Spectrum Protect storage.
If the files are needed, they must be manually retrieved back to the file system. A variation of
this is active archiving, which is a mechanism for moving data between different tiers of
storage depending on its retention requirements. For example, data that is in constant use is
kept on high-performance disk drives, and data that is rarely referenced or is required for
long-term retention is moved to lower performance disk or tape drives.
36 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
IBM Spectrum Archive EE uses the third option, which instead of deleting the archived files, it
creates a stub file in its place. If the files are needed, they are automatically retrieved back to
the file system by using the information that is stored in the stub file when they are accessed.
The final type of archiving is compliant archiving, which is a legislative requirement of various
countries and companies data retention laws, such as Sarbanes-Oxley in the US. These laws
require a business to retain key business information. Failure to comply with these laws can
result in fines and sanctions. Essentially, this type of archiving results in data being stored by
the backup software without the possibility of it being deleted before a defined period elapses.
In certain cases, it can never be deleted.
The most current information for IBM Spectrum Archive EE hardware and software
configurations, notices, and limitations can always be found in the readme file of the software
package.
IBM Spectrum Archive Enterprise Edition is to be installed and configured on one or more
IBM Spectrum Scale nodes. IBM Spectrum Archive EE installed nodes make up an IBM
Spectrum Archive EE cluster within an IBM Spectrum Scale cluster. The IBM Spectrum
Archive EE nodes need to be attached to a supported tape library facility. A maximum of two
tape libraries are supported for an IBM Spectrum Archive EE cluster.
From an IBM Spectrum Scale architecture perspective, the IBM Spectrum Archive EE
software-defined storage can be installed and configured on an IBM Spectrum Scale Client
node, an IBM Spectrum Scale Server node, or an IBM Spectrum Scale Protocol node. The
node that you choose must have the supported requirements for IBM Spectrum Archive EE.
Note: IBM Spectrum Archive EE can NOT be installed on IBM Elastic Storage® Server
(IBM ESS) nodes. See 3.3.1, “Limitations” on page 51.
This section describes some of the typical deployment configurations for IBM Spectrum
Archive EE and provides insights to help plan for its implementation.
40 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-1 shows the physical and its equivalent logical diagram that show the minimum
architecture approach for IBM Spectrum Archive EE.
The required software components of IBM Spectrum Scale and IBM Spectrum Archive EE
are installed in one physical server either directly Fibre Channel-attached to an IBM tape
library or using a storage area network. The RAID-protected LUNs and IBM tape library are
configured in IBM Spectrum Scale and IBM Spectrum Archive EE as pools - flash pools,
spinning disk pools and tape pools. This architecture deployment option enables automated
data tiering based on user-defined policies with minimum solution components.
This architecture is also known as Long-Term Archive Retention (LTAR). However, as this
solution is usually deployed for long term data archive retention, injecting high availability is
highly recommended.
Figure 3-2 High availability approach IBM Spectrum Archive EE on IBM Spectrum Scale Servers
Two physical servers are deployed each with IBM Spectrum Scale and IBM Spectrum Archive
EE software components. Both servers are connected to a storage area network which
provides connectivity to a shared storage area network (SAN) storage system and an IBM
tape library. This approach ensures that both servers are connected to the storage systems
by way of a high-speed Fibre Channel network. The connectivity of the cluster to a shared
flash or disk storage system provides high performance and high availability in the event of a
server outage. For more information about configuring a multiple-node cluster, refer to 5.2.3,
“Configuring a multiple-node cluster” on page 112.
Another approach to implementing high availability for data on tapes is the deployment of two
or more tape pools. Policies can be configured to create up to three copies of data in the
additional tape pools.
Note: Multiple tape pools also can be deployed on the minimum solution, as shown in
Figure 3-1 on page 41.
42 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3.1.2 As an IBM Spectrum Scale Client
On this architecture, the IBM Spectrum Archive EE node is deployed as a separate
software-defined storage module integrating with the IBM Spectrum Scale single namespace
file system with its own Storage Area Network interface to the tape platform.
Figure 3-3 shows this unique approach that leverages software-defined storage for flexibility
of scaling the system.
Figure 3-3 IBM Spectrum Archive EE node as an IBM Spectrum Scale Client
The diagram shows a single node IBM Spectrum Archive EE as an IBM Spectrum Scale
Client with a single tape library attached using SAN. However, it is easy to add IBM Spectrum
Archive EE nodes that are attached to the same library using SAN for high-availability or
higher tape-access throughput.
This is the solution approach for customers with IBM Spectrum Scale environments who want
to use tapes to stream line costs for long-term archive retention. This architecture provides a
platform to scale both IBM Spectrum Scale and IBM Spectrum Archive EE independently, that
is by adding nodes to each of the clusters when needed. The integration of the server nodes
with the storage systems can be done over the same SAN fabric considering the appropriate
SAN zoning implementations. This way, scaling the IBM Spectrum Scale flash- or
disk-storage capacity is accomplished by merely adding flash or disk drives to the storage
server.
Figure 3-4 IBM Spectrum Archive EE node as an IBM Spectrum Scale Client in an IBM ESS
This architecture allows the same easy scalability feature as the IBM Spectrum Scale Client
described in 3.1.2, “As an IBM Spectrum Scale Client” on page 43; With this scalability
feature, you can add nodes when needed. In Figure 3-4, you can add IBM ESS modules to
expand the IBM Spectrum Scale cluster; or you can add IBM Spectrum Archive EE nodes to
expand the file system tape facility channels. Needless to say, adding tape drives and tape
cartridges to address growing tape access and storage requirements is also simple and
nondisruptive.
44 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-5 shows a single namespace file system IBM Spectrum Scale stretched cluster
architecture deployed across two sites.
Figure 3-5 IBM Spectrum Scale Stretched Cluster across metropolitan distances of up to 300 KM
This architecture shows a good use case for deploying two tape libraries in one IBM
Spectrum Scale cluster. For sites that are geographically separated by greater than 300 KM,
Active File Management (AFM) might be deployed to replicate data across the sites. However,
both sites are independent IBM Spectrum Scale clusters as compared to the single
namespace file system over a stretched cluster.
For more information about stretched clusters, see 5.2.4, “Configuring a multiple-node cluster
with two tape libraries” on page 115.
For example use cases on stretched clusters and AFM, see 8.8, “University Scientific Data
Archive” on page 286 and 8.11, “AFM use cases” on page 290.
Note: Figure 3-5 shows the IBM Spectrum Scale cluster deployed as an IBM ESS
appliance. The cluster might also be a software-defined storage deployment on IBM
Spectrum Scale Server nodes as discussed in 3.1.1, “On IBM Spectrum Scale Servers” on
page 40 and 3.1.2, “As an IBM Spectrum Scale Client” on page 43.
There are three ways of accessing data on an IBM Spectrum Archive EE deployed IBM
Spectrum Scale system:
3.2.1 Data access using application or users on IBM Spectrum Scale Clients
This access method refers to compute nodes directly accessing the IBM Spectrum Scale
System. Each compute node on which the applications are running, needs to have an IBM
Spectrum Scale Client installed for high-speed data access to the IBM Spectrum Scale
Cluster.
Figure 3-6 shows that the IBM Spectrum Scale Client is installed on the compute nodes
where the applications are running.
Figure 3-6 Access using IBM Spectrum Scale Client on compute node
Note: Each compute node must have IBM Spectrum Scale Clients installed to access the
IBM Spectrum Scale Cluster. This allows high-performance (network dependent) access to
the cluster.
46 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3.2.2 Data access using Protocol Nodes on IBM Spectrum Scale Server Nodes
with IBM Spectrum Archive EE
This access method can be more “application-friendly” because alterations or software
modules are not required to be installed on the application compute-nodes. Access to the IBM
Spectrum Scale Cluster is gained by using the Protocol Nodes which offers other choices of
data interface. The following data interfaces might be used by the application compute-nodes
through the IBM Spectrum Scale Protocol Nodes:
Network File System (NFS)
Server Message Block (SMB)
Hadoop Distributed File System (HDFS)
Object
Figure 3-7 shows this architecture where Protocol Nodes are deployed on the same servers
as that of the IBM Spectrum Scale Server and IBM Spectrum Archive EE nodes.
Figure 3-7 Access using Protocol Nodes on IBM Spectrum Scale and IBM Spectrum Archive EE nodes
Figure 3-8 shows another version of this access method, where Application Compute Nodes
access the IBM Spectrum Scale Cluster using Protocol Nodes. However, this architecture
deploys the Protocol Nodes in separate hosts or virtual machines. This way, scaling is much
simpler, that is, you can add the required nodes when needed.
48 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3.2.3 Data access using Protocol Nodes integrated with the IBM ESS
This access method is similar to Figure 3-8. However, instead of software-defined storage
modules implemented on servers or hosts, the IBM Spectrum Scale system is deployed on an
IBM ESS environment. The IBM Spectrum Archive EE environment is also deployed
independently as an IBM Spectrum Scale Client, which provides flexible scalability for all
components.
Figure 3-9 shows an IBM ESS system integrated with an IBM Spectrum Archive EE node. On
this architecture, a separate cluster of Protocol Nodes are deployed to provide applications
and users with access to the IBM Spectrum Scale in the IBM ESS system.
This architecture uses the pre-tested and pre-installed IBM Spectrum Scale environment as
an IBM ESS appliance. Scaling this architecture is simplified, where upgrades can be
deployed to the independent clusters (which make up the IBM Spectrum Scale single
namespace filesystem). Also, note that IBM Spectrum Archive EE nodes can be added to this
architecture for high availability or to increase tape access performance.
Supported operating Red Hat Enterprise Linux Server 7.4, 7.5, 7.6, 7.7, 7.8, 7.9, 8.2, 8.3, and 8.4
systems (x86_64)
Supported operating Red Hat Enterprise Linux Server 7.4, 7.5, 7.6, 7.7, 7.8, 7.9, 8.2, 8.3, and 8.4
systems (ppc64le)
Supported tape libraries IBM Spectrum Archive Enterprise Edition supports up to two tape libraries of the following
types:
IBM TS4500 tape library (IBM LTO 5 and later generation LTO tape drives, TS1140
and later generation 3592 tape drives)
IBM TS4300 tape library (IBM LTO 6 and later generation LTO tape drives)
IBM TS3500 tape library (IBM LTO 5 through LTO 8 generation LTO tape drives,
TS1140 and later generation 3592 tape drives)
IBM TS3310 tape library (IBM LTO 5 and later generation LTO tape drives)
LTO media: LTO 9, LTO 8, LTO 7 Type M8, LTO 7, LTO 6, and LTO 5
Recommended Access IBM TS1140, IBM TS1150, IBM TS1155, IBM TS1160, and LTO-9
Order (RAO) support
Server
50 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Linux computers
Memory Minimum: 2 x (d) x (f) + 1 GB of RAM available for the IBM Spectrum Archive EE
program:
– d: Number of tape drives
– f: Number of millions of files/directories on each tape cartridge in the system
In addition, IBM Spectrum Scale must be configured with adequate RAM.
Example: There are six tape drives in the system and three million files are stored on
each tape cartridge. The minimum required RAM is 37 GB (2 x 6 x 3 + 1 = 37).
Preferred: 64 GB RAM and greater
Host Bus Adapter (HBA)a, Minimum: Fibre Channel Host Bus Adapter supported by TS1160, TS1155, TS1150,
RoCE TS1140, LTO-9, LTO-8, LTO-7, LTO-6, and LTO-5 tape drives
Preferred: 8 Gbps/ 16 Gbps Dual port or Quad port Fibre Channel Host Bus Adapter
Disk device for LTFS EE For more information, see 3.7, “Sizing and settings” on page 61.
tape file system metadata
One or more disk devices The amount of disk space that is required depends on the IBM Spectrum Scale settings
for the GPFS file system that are used.
a. For more information about HBA interoperability, see the IBM System Storage Interoperation Center (SSIC) web
page.
3.3.1 Limitations
In this section, we describe the limitations of IBM Spectrum Archive EE.
The hard link creates an alias for the data with different name or directory path without
copying it. The original file and other hard links all point to the same data on disk.The view
from other hard links is affected if the file that is linked from one hard link is changed.
If multiple hard links are necessary on the IBM Spectrum Archive EE managed file system,
those files must be excluded from policy files; otherwise, commands can have unexpected
results for the user. The following consequences can occur on running the commands:
Migration, Recall, and Premigrate commands
If one of the hard links is specified in a migration task, all hard links of that file are shown
as migrated. If multiple hard links are specified in a single migration task, the migration
results in a “duplicate” error. If one of the hard links is specified in a recall task, all hard
links of that file are shown as recalled. If multiple hard links are specified in a single recall
task, the recall results in a “duplicate error”.
Save commands
The hard link information is not saved with the save command because the objects contain
data.
File behavior on tape export and imports
The tape export process can cause unintentional unlinks from I-nodes. Even if all hard
links indicate that they are migrated to tape, multiple hard links for a single I-node is not
created during the tape import.
52 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Limitations on IBM Spectrum Scale features
The file system that is managed by IBM Spectrum Archive EE cannot be used with the
following functions or features of IBM Spectrum Scale:
Scale-out Backup and Recovery (SOBAR) for DR purposes
Advanced File Management (AFM), in a mode other than Independent Writer mode
Transparent Cloud Tiering (TCT)
Remote Mount
IBM Spectrum Archive EE handles only file systems that belong to the local (home) IBM
Spectrum Scale cluster, but not file systems that are installed remotely.
3.4.1 Required software packages for Red Hat Enterprise Linux systems
The following software modules are required to deploy IBM Spectrum Archive Enterprise
Edition on Red Hat Enterprise Linux systems:
The most current fix pack for IBM Spectrum Scale: 5.0.2, 5.0.3, 5.0.4, 5.0.5, 5.1.0, 5.1.1,
5.1.2, or subsequent releases. The following operating system software:
– attr
– Boost.Date_Time (boost-date-time)
– Boost.Filesystem (boost-filesystem)
– Boost.Program_options (boost-program-options)
– Boost.Serialization (boost-serialization)
– Boost.Thread (boost-thread)
– boost_regex
– FUSE
– fuse
– fuse-libs
– gperftools-libs
– Java virtual machine (JVM)
– libxml2
– libuuid
– libicu
– lsof
– net-snmp
– nss-softokn-freebl
– openssl
– Python 2.4 or later, but earlier than 3.0
– python3-pyxattr (required with Red Hat Enterprise Linux 8.x. Requires access to the
Red Hat CodeReady Linux Builder repository)
– pyxattr (required with Red Hat Enterprise Linux 7.x)
– rpcbind
– sqlite
HBA device driver (if one is provided by the HBA manufacturer) for the host bus adapter
connecting to the tape library and attach to the tape drives.
54 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Note: Java virtual machine (JVM) 1.7 or later must be installed before the “Extracting
binary rpm files from an installation package” step for IBM Spectrum Archive on a Linux
system during the installation of IBM Spectrum Archive Enterprise Edition.
Note: The REST API must be installed on one of the IBM Spectrum Archive Enterprise
Edition nodes.
Table 3-2 Valid combinations for types of nodes in the IBM Spectrum Scale cluster
Node type IBM IBM Spectrum IBM Multi-tape
Spectrum Archive internal Spectrum management
Scale Hierarchical Archive LE module (MMM)
Storage
Management (HSM)
All other combinations are invalid as an IBM Spectrum Archive EE system. IBM Spectrum
Archive EE nodes have connections to the IBM tape libraries and drives.
Multiple IBM Spectrum Archive EE nodes enable access to the same set of IBM Spectrum
Archive EE tapes. The purpose of enabling this capability is to increase the performance of
the host migrations and recalls by assigning fewer tape drives to each IBM Spectrum Archive
EE node. The number of drives per node depends on the HBA/switch/host combination. The
idea is to have the maximum number of drives on the node such that all drives on the node
can be writing or reading at their maximum speeds.
A single NUMA node is preferable for better performance. For servers that contain multiple
CPUs, the key is to remove memory from the other CPUs and group them in a single CPU to
create a single NUMA node. This configuration allows all the CPUs to access the shared
memory, resulting in higher read/write performances between the disk storage and tape
storage.
FC switches can be added between the host and tape drives and between the host and the
disk storage to create a storage area network (SAN) to further expand storage needs as
required.
56 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3.6 IBM Spectrum Archive deployment examples
This section describes some examples of IBM Spectrum Archive deployment on Lenovo
severs and on a Versastack converged infrastructure.
Figure 3-10 Minimum deployment on Lenovo ThinkSystem SR650 server with IBM TS4300
The following example server configuration is used for a Lenovo ThinkSystem SR650 rack
server:
Forty cores (Two sockets each with 20-cores) Intel Xeon Processors
512 GB RAM
RAID controller
Nine units 10 TB 3.5-inch 7.2KRPM NL SAS (RAID 6 = 6D+P+Q+Spare)
Five units 3.84 TB 3.5-inch SATA SSD (RAID 5 = 3D+P+Spare)
Two units 480 GB SATA Non-Hot Swap SSD for operating system
Red Hat Enterprise Linux with Lenovo Support for Virtual Data Centers x2 socket licenses
One unit Dual-port 16 Gbit FC HBA for tape interface
One unit Dual-port 10/25 GbE SFP28 PCIe Ethernet adapter for network interface
One available PCIe slot might be configured as other FC or GbE connectivity
IBM Spectrum Scale can be licensed on a per-terabyte basis with the following three license
options:
Data Access Edition
Data Management Edition
Erasure Code Edition.
The storage capacity to be licensed is the capacity in terabytes (TiB) from the network shared
disk (NSD) in the IBM Spectrum Scale Cluster. This means that less capacity can be
allocated to IBM Spectrum Scale if the required capacity of the storage tier required is less
than 65 TiB, thereby reducing the IBM Spectrum Scale capacity license.
For more information about IBM Spectrum Scale licenses and licensing models, see the
following IBM Documentation web pages:
IBM Spectrum Scale license designation
Capacity-based licensing
The IBM TS4300 tape library is configured with two units of half-high LTO-8 tape drives, which
directly connect to the Lenovo ThinkSystem SR650 server using 8Gbit Fibre-Channel
connections.
IBM Spectrum Archive EE is licensed per node regardless of the capacity migrated to tapes.
With node-based licensing, storing data on tapes is virtually unlimited provided if enough tape
resources are available in the tape-storage tier. This node-based licensing can help realize
significant savings on storage and operational costs.
In this example, we need only one IBM Spectrum Archive EE license, which is for one node
that is installed with IBM Spectrum Scale. If more tape access channels are needed, more
IBM Spectrum Scale with IBM Spectrum Archive EE nodes can be added to the cluster.
58 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-11 High availability on minimum deployment IBM Spectrum Scale
The difference of this architecture is that instead of internal storage for the IBM Spectrum
Scale with IBM Spectrum Archive EE servers, storage capacity will now be on a shared
storage system using redundant SAN Switches. In Figure 3-11, the IBM FlashSystem®
FS5035 is deployed with a pair of IBM SAN48-C SAN switches for this example.
The IBM Spectrum Scale license for this architecture can be either Data Access Edition or
Data Management Edition. Erasure Code Edition is not applicable because of the shared
storage architecture. In Figure 3-11, both servers with IBM Spectrum Archive EE need to be
licensed.
60 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-12 shows a high-level overview of an IBM Spectrum Archive EE architecture
deployed on Versastack.
Note: You must collaborate with your in-country Cisco Data Center Product Specialists for
the sizing of the compute, network, and services requirements.
The IBM Spectrum Archive EE metadata directory can be stored in its own IBM Spectrum
Scale file system or it can share the IBM Spectrum Scale file system that is being
space-managed.
The size requirements of the IBM Spectrum Scale file system that is used to store the IBM
Spectrum Archive EE metadata directory depends on the block size and the number of files
that are migrated to IBM Spectrum Archive EE.
The following calculation produces an estimate of the minimum number of inodes that the
IBM Spectrum Scale file system must have available. The calculation depends on the number
of cartridges:
Number of inodes = 500 + (15 x c) (Where c is the number of cartridges.)
Important: If there is more than one tape library, the number of cartridges in your
calculation must be the total number of cartridges in all libraries.
The following calculation produces an estimate of the size of the metadata that the IBM
Spectrum Scale file system must have available:
Number of GBs = 10 + (3 x F x N)
where:
F is the number of files, in millions, to migrate.
N is the number of replicas to create.
For example, to migrate 50 million files to two tape storage pools, 310 GB of metadata is
required:
10 + (3 x 50 x 2) = 310 GB
Thus, to ensure that file migration can occur on redundant copy tapes, the number of tapes in
the redundant copy pools must be the same or greater than the number of tapes in the
primary copy pool. For example, if the primary copy pool has 10 tapes, the redundant copy
pools also should have at least 10 tapes. For more information about redundant copies, see
6.11.4, “Replicas and redundant copies” on page 177.
Note: The most commons setup is to have two copies, that is, one primary and one copy
pool with the same number of tapes.
62 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3.7.2 Planning for LTO-9 Media Initialization/Optimization
The deployment of LTO-9 tape drives will require a one-time media optimization for all new
LTO-9 tape cartridges. Because the process of media optimization can take an average of 40
minutes up to two hours per LTO-9 tape cartridge, refer to the suggested strategies to mitigate
disruptions to normal tape operations:
Small tape library deployments are recommended to perform media optimization to all
media before beginning normal operations.
Large tape library deployments are recommended to use as many drives to perform media
optimization as possible. Then dedicate one drive for every seventeen drives to
continuously load new media for optimization during normal operations.
For more information about LTO-9 media initialization and optimization, see 7.29, “LTO 9
Media Optimization” on page 272, and this Ultrium LTO web page.
Note: The media optimization process cannot be interrupted; otherwise, the process must
restart from the beginning.
Before simulating configurations with the IBM Spectrum Archive EE Sizing Calculator, the
workload data-characteristics first must be studied.
Understanding the workload data file sizes is key to the designing of the IBM Spectrum Scale
file system. The file size of most of the data determines the suitable Data and Metadata block
sizes during initial IBM Spectrum Scale deployment for new systems.
For more information about Data and Metadata, see IBM Spectrum Scale Version 5.1.2
Concepts, Planning, and Installation Guide.
For small to medium implementations like the architecture depicted in figure 3.1.1, “On IBM
Spectrum Scale Servers” on page 40, it is recommended to consider flash storage for the
System Pool (which is where the Metadata is stored in the file system) for performance. More
Storage Pools can be designed by using flash or spinning disks depending on data tiering
requirements.
64 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Note: For converged solution architectures where IBM Spectrum Scale and IBM Spectrum
Archive EE are installed in one physical server, it is recommended to assign at least
128GB of memory for the IBM Spectrum Archive EE and respectively add the IBM
Spectrum Scale requirements that will be deployed on the same server. Apply the same for
CPU and connectivity requirements.
3.7.4 Performance
Performance planning is an important aspect of IBM Spectrum Archive EE implementation,
specifically migration performance. The migration performance is the rate at which IBM
Spectrum Archive EE can move data from disk to tape, freeing up space on disk. The number
of tape drives (including tape drive generation) and servers that are required for the
configuration can be determined based on the amount of data that needs to be moved per
day and an estimate of the average file size of that data.
Note: Several components of the reference architecture affect the overall migration
performance of the solution, including backend disk speeds, SAN connectivity, NUMA
node configuration, and amount of memory. Thus, this migration performance data should
be used as a guideline only and any final migration performance measurements should be
done on the actual customer hardware.
For more migration performance information, see IBM Spectrum Archive Enterprise Edition
v1.2.2 Performance.
x8
x6
QLogic® QLE2562
8Gb FC HBA.
x8 x6
QLogic® QLE2562
8Gb FC HBA.
The performance data shown in this section was derived by using two x3850 X5 servers
consisting of multiple QLogic QLE2562 8 Gb FC HBA cards. The servers were also modified
by moving all the RAM memory to a single CPU socket to create a multi-CPU, single NUMA
node. The HBAs were relocated so they are on one NUMA node.
This modification was made so that memory can be shared locally to improve performance.
The switch used in the lab was an IBM 2498 model B40 8 Gb FC switch that was zoned out
so that it had a zone for each HBA in each node. An IBM Storwize® V7000 disk storage unit
was used for the disk space and either TS1150 or TS1070 drives were used in a TS4500 tape
library.
Figure 3-13 shows the example configuration. This configuration is only one possibility; yours
might be different.
66 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-14 shows a basic internal configuration of Figure 3-13 in more detail. The two x3850
X5 servers are using a single NUMA and have all the HBAs performing out of it. This
configuration allows all CPUs within the servers to work more efficiently. The 8 GB Fibre
Channel switch is broken up so that each zone handles a single HBA on each server.
Storage
TS4500
Note: Figure 3-14 shows a single Fibre Channel cable going to the drive zones. However,
generally you should have a second Fibre Channel cable for failover scenarios. This is the
same reason why the zone that goes to the external storage unit has two Fibre Channel
cables from 1 HBA on each server.
Figure 3-14 is one of many possible configurations and can be used as a guide for how you
set up your own environment for best performances. For example, if you have more drives,
you can add tape drives per zone, or add Host bus adapters (HBAs) on each server and
create a zone.
With a TS1150 configuration of 1N4D4T, you can expect to see a combined total transfer rate
of 1244.9 MBps for 10 GiB files. If that configuration is doubled to 2N4D8T, the total combined
transfer rate is nearly doubled to 2315.3 MiB/s for 10 GiB files. With this information, you can
estimate the total combined transfer rate for your configuration.
Table 3-3 TS1150 raw performance data for multiple node/drive configurations with 5 MiB, 10 MiB, 1 GiB, and 10 GiB files
Node/Drive File size
configuration
5 MiB 10 Mib 100 MiB 1 Gib 10 Gib
Table 3-4 lists the raw performance for the LTO7 drives.
Table 3-4 LTO7 raw performance data for multiple node/drive configurations with 5 MiB, 10 MiB, 1 GiB, and 10 GiB files
Node/Drive File size
configuration
5 MiB 10 Mib 100 MiB 1 Gib 10 Gib
68 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 3-15 shows a comparison line graph of the raw performance data obtain for TS1150
drives and LTO7 drives that use the same configurations.
2000.0 2016.4
8 Tape Drives (LTO7)
1921.7
1731.8
TRANSFER RATE (MB/SEC)
1500.0
1473.4
624.3
583.6
2 Tape Drives (TS1150)
577.3
465.8
500.0 561.8 2 Tape Drives (LTO7)
493.7
369.0 439.2
406.7
365.4 186.3
114.0
178.4
0.0 111.0
1 10 100 1000 10000 100000
FILE SIZE (MEGA BYTES)
Figure 3-15 Comparison between TS1150 and LTO-7 drives using multiple node/drive configurations
It is clear that by using small files, the comparison between the two types of drives is minimal.
However, when migrating file sizes of 1 GiB and larger, a noticeable difference exists.
Comparing the biggest configuration of 2N4D8T, the LTO peaks at a total combined transfer
rate of 1921.7 MiBps. With the same configuration but with TS1150 drives, it peaks at a total
combined transfer rate of 2315.3 MiBps.
The range from 7620 to (7630 + The IBM Spectrum Archive EE MMM
number of EE nodes) active control node
RPC bind daemon port (for The IBM Spectrum Archive EE MMM
example, 111) active control node
Several other ports are used by HSM. For more information, see this IBM Documentation web
page.
IBM Spectrum Scale disk Method 1. Preserve GPFS cluster and GPFS file systems
storage (internal and external) Follow IBM Spectrum Scale procedure for NSD add, NSD restripe and NSD
delete. For more information, see:
– NSD add
– NSD restripe
– NSD delete
70 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Hardware upgrade Upgrade procedure
Tape drives/cartridges Method 1. IBM Spectrum Archive EE eeadm pool set command to modify the
technology upgrade media_restriction attribute
Use IBM Spectrum Archive media_restriction attribute of a pool to gradually
move data to the specified newer tape technology within a pool.
For more information, see 6.11.5, “Data Migration” on page 180 for details.
Tape library For more information, see 4.6, “Library replacement” on page 88.
Although IBM Spectrum Archive EE is based on the IBM Linear Tape File System standard
components as IBM Spectrum Archive and the IBM Spectrum Archive Library Edition (LE),
these components are all included with the IBM Spectrum Archive EE installation package.
Before you can start with the installation routines, you must verify the following installation
prerequisites:
Installation prerequisites (see 4.2, “Installation prerequisites for IBM Spectrum Archive
EE” on page 74)
This section describes the tasks that must be completed before installing IBM Spectrum
Archive EE.
Installing IBM Spectrum Archive EE on a Linux server (see 4.3, “Installing IBM Spectrum
Archive EE” on page 75)
This section describes how to install the IBM Spectrum Archive EE package on a Linux
server.
Ensure that the following prerequisites are met before IBM Spectrum Archive EE is installed.
For more information, see the other topics in this section if needed.
Verify that your computer meets the minimum hardware and software requirements for
installing the product. For more information, see Chapter 3, “Planning for IBM Spectrum
Archive Enterprise Edition” on page 39.
Verify that your user ID meets the requirements for installing the product (such as you are
working with the root user ID or have the root administration permissions).
Ensure that you reviewed all of the planning information that is described in Chapter 3,
“Planning for IBM Spectrum Archive Enterprise Edition” on page 39.
If the standard IBM Spectrum Archive LE is already installed, it must be uninstalled before
IBM Spectrum Archive EE is installed. For more information, see 10.4, “IBM Spectrum
Archive EE interoperability with IBM Spectrum Archive products” on page 330.
Ensure that all prerequisite software is installed, as described in 4.2.1, “Installing the host
bus adapter and device driver” on page 75.
Ensure that the host bus adapter (HBA) and device driver are installed, as described in
4.2.1, “Installing the host bus adapter and device driver” on page 75.
74 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Determine the distribution package for IBM Spectrum Archive EE that is required for your
system.
For IBM Spectrum Scale prerequisites, see 3.5, “Hardware and software setup” on
page 56.
To install the HBA and its device driver, see the documentation that is provided by the HBA
manufacturer.
If the HBA attached to the tape library is an Emulex adapter, add the following line to the
/etc/modprobe.d/lpfc.conf file:
options lpfc lpfc_sg_seg_cnt=256
Then, restart the server system for the change to take effect.
For more information about fixes and updates for your system’s software, hardware, and
operating system, see IBM Support’s Fix Central web page.
For more information about HBA interoperability, see this IBM System Storage Interoperation
Center (SSIC) web page.
Interactive console mode is the method that is used for extracting binary rpm files from an
installation package.
Before you use any of these methods to extract the IBM Spectrum Archive EE binary rpm
files, you must confirm or set the run permission of the installation package.
Before the IBM Spectrum Archive EE binary rpm files are extracted from an installation
package, complete the following steps:
1. Confirm the run permission of ltfsee-[version]-[buildlevel].bin by running the
following command:
# ls -l ltfsee-[version]-[buildlevel].bin
2. If it is not already set, set the run permission by running the following command:
# chmod +x ltfsee-[version]-[buildlevel].bin
3. Proceed with the extraction of the binary IBM Spectrum Archive EE rpm files by selecting
one of the procedures that are described next.
In the lab setup that was used for this book, we used the interactive console mode method,
which is the option most users are likely to use.
Important: The steps in this section extract binary rpm files to your local disk only. To
complete the installation process, more steps are required. After you complete the
extraction of the binary rpm files, see “Installing, upgrading, or uninstalling IBM Spectrum
Archive EE automatically” on page 79 or 4.5, “Quick installation guide for IBM Spectrum
Archive EE” on page 88 for more information.
To extract IBM Spectrum Archive EE binary rpm files in interactive console mode, complete
the following steps:
1. Run IBM Spectrum Archive EE installation package on the system by running the
appropriate command for your environment:
– If your operating system is running on the command-line interface (CLI), run the
following command:
#./ltfsee-[version]-[buildlevel].bin
– If your operating system is running on the GUI (X Window System), run the following
command:
#./ltfsee-[version]-[buildlevel].bin -i console
76 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The messages that are shown in Example 4-1 are displayed.
Launching installer...
Important: You cannot select the installation and link folders with the console installer.
They are created in the ~/LTFSEE/ directory, which is the default folder of the installer
that extracts the required files.
The installation script ltfsee_install for the command-line installation is found under
the ~/LTFSEE/rpm.[version]_[buildlevel] folder; for example, the
~/LTFSEE/rpm.1322_52823/ subfolder.
2. Read the International Program License Agreement. Enter 1 to accept the agreement and
press Enter to continue, as shown in Example 4-2.
===============================================================================
1. Definitions
[==================|==================|==================|==================]
[------------------|------------------|------------------|------------------]
When the files are successfully extracted, the text-based installer completes.
Important: The following symbolic links are created in your home directory:
A link to the rpm folder that keeps the extracted rpm files.
A link to the “Change IBM Linear Tape File System Enterprise Edition Installation”
executable file that uninstalls IBM Spectrum Archive EE.
When you successfully finish, continue to “Installing, upgrading, or uninstalling IBM Spectrum
Archive EE automatically” on page 79 to complete the installation. If you prefer to install
manually, see 4.5, “Quick installation guide for IBM Spectrum Archive EE” on page 88.
IBM Spectrum Archive EE nodes communicate by using several TCP/UDP ports. Because
some ports are assigned dynamically within a wide range, you must disable any firewall
program that runs on these nodes.
78 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Important: On Red Hat Enterprise Linux 7 and 8, managing system services are done by
using the systemctl command.
To stop the firewall service, run the systemctl stop firewalld command.
To prevent the firewall service from being automatically started at start time, run the
systemctl disable firewalld command.
In addition, you can mask the firewalld service to prevent it from being started manually or
by another service by running the systemctl mask firewalld command.
During this installation for the IBM Spectrum Archive EE rpm files, there is also an MIB file
that is provided if you plan to use SNMP for monitoring of your IBM Spectrum Archive EE
setup. SNMP monitoring software usually requires such an MIB file to manage the various
SNMP traps sent to it. The IBM Spectrum Archive EE MIB file is in the
/opt/ibm/ltfsee/share/IBMSA-MIB.txt directory.
The automated method is based on a utility (a shell script), which is provided by the IBM
Spectrum Archive EE installation package. The script is named ltfsee_install and can be
found after extracting the binary installation files in the
~/LTFSEE/rpm.[version]_[buildlevel] directory with the IBM Spectrum Archive EE rpm files
(such as /root/LTFSEE/rpm.1322_52823/).
ltfsee_install utility
Use the ltfsee_install command-line utility to install rpm packages automatically to the
system. You must have root user authority to use this command.
For more information, see “Installing, upgrading, or uninstalling IBM Spectrum Archive EE
automatically” on page 79.
The ltfsee_install <option> command installs the following rpm packages to the system:
IBM Spectrum Archive LE component
Integrated customized IBM Spectrum Protect for Space Management with IBM Spectrum
Archive EE
IBM Spectrum Archive Migration Driver
Verify that these conditions are met by logging on to the operating system as a root user and
running the following command:
# ./ltfsee_install --check
If the conditions are met, the following message is shown as the last line of output:
The prerequisites checking is completed successfully.
The ltfsee_install file installs or upgrades all required rpm packages on the server node. It
can also uninstall those rpm packages from the node if needed.
Complete the following steps to automatically install, upgrade, or uninstall IBM Spectrum
Archive EE by running the ltfsee_install command:
1. Log on to the operating system as a root user.
2. On each node in the cluster, complete the set of tasks for the action you want to take:
a. Installing IBM Spectrum Archive EE on the node:
i. Run the following command:
# ./ltfsee_install --install
Example 4-5 shows you the complete output of the ltfsee_install --install
command.
ii. Verify that the command completed successfully. Check for the following success
message in the command output:
All rpm packages are installed successfully.
80 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Installing HSM modules.
Created symlink from
/etc/systemd/system/multi-user.target.wants/hsm.service to
/usr/lib/systemd/system/hsm.service.
Installing EE modules.
Starting HSM.
Completed RPM installation.
Running the post-installation steps.
All rpm packages are installed successfully.
The installed IBM Spectrum Archive EE version: 1.3.2.0_52814.
Complete the configuration by using the /opt/ibm/ltfsee/bin/ltfsee_config
command.
** ATTENTION **
For problem determination, it is strongly recommended that you disable
log suppression and set up the abrtd daemon to capture the Spectrum
Archive core dumps.
Refer to the Troubleshooting section of the IBM Spectrum Archive
Enterprise Edition documentation in the IBM Knowledge Center.
The following files were updated during the installation. (The original
files were copied to the /opt/ibm/ltfsee/bin directory)
- /etc/rsyslog.d/ibmsa-rsyslog.conf
- /etc/logrotate.d/ibmsa-logrotate
iv. Verify that the command completed. Check for the following success message in
the command output:
All rpm packages are installed successfully.
c. Uninstalling IBM Spectrum Archive EE from the node:
i. Run the following command:
# ltfsee_install --clean
ii. Verify that the command completed. Check for the following success message in
the command output:
Uninstallation is completed.
3. Verify that the installation or uninstallation completed successfully by running the following
command:
# ltfsee_install --verify
If the installation (see Example 4-7) was successful, the following message is shown:
Module installation check is completed.
82 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
An IBM Spectrum Archive EE package ltfsle-library is installed.
An IBM Spectrum Archive EE package ltfsle-library-plus is installed.
An IBM Spectrum Archive EE package ltfsle is installed.
An IBM Spectrum Archive EE package ltfs-admin-utils is installed.
An IBM Spectrum Archive EE package gskcrypt64 is installed.
An IBM Spectrum Archive EE package gskssl64 is installed.
An IBM Spectrum Archive EE package TIVsm-API64 is installed.
An IBM Spectrum Archive EE package TIVsm-BA is installed.
An IBM Spectrum Archive EE package TIVsm-HSM is installed.
An IBM Spectrum Archive EE package ltfs-mig is installed.
All IBM Spectrum Archive EE packages are installed.
The rest service can be installed on any node within the cluster that has IBM Spectrum
Archive EE installed. To start the installation, some software requirements must be met before
the rpm can be installed.
Example 4-8 shows how to perform the installation of the required software to run REST.
Dependencies Resolved
==================================================================================
===================================
Package Arch Version
Repository Size
==================================================================================
===================================
Installing:
httpd x86_64 2.4.6-45.el7
rhel-x86_64-server-7 1.2 M
mod_ssl x86_64 1:2.4.6-45.el7
rhel-x86_64-server-7 105 k
mod_wsgi x86_64 3.4-12.el7_0
rhel-x86_64-server-7 76 k
Installing for dependencies:
httpd-tools x86_64 2.4.6-45.el7
rhel-x86_64-server-7 84 k
mailcap noarch 2.1.41-2.el7
rhel-x86_64-server-7 31 k
Transaction Summary
==================================================================================
===================================
Install 3 Packages (+2 Dependent packages)
84 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Installing : httpd-2.4.6-45.el7.x86_64
3/5
Installing : mod_wsgi-3.4-12.el7_0.x86_64
4/5
Installing : 1:mod_ssl-2.4.6-45.el7.x86_64
5/5
Verifying : httpd-tools-2.4.6-45.el7.x86_64
1/5
Verifying : mod_wsgi-3.4-12.el7_0.x86_64
2/5
Verifying : mailcap-2.1.41-2.el7.noarch
3/5
Verifying : 1:mod_ssl-2.4.6-45.el7.x86_64
4/5
Verifying : httpd-2.4.6-45.el7.x86_64
5/5
Installed:
httpd.x86_64 0:2.4.6-45.el7 mod_ssl.x86_64 1:2.4.6-45.el7
mod_wsgi.x86_64 0:3.4-12.el7_0
Dependency Installed:
httpd-tools.x86_64 0:2.4.6-45.el7 mailcap.noarch
0:2.1.41-2.el7
Complete!
If pip is not installed on the designated node (pip is installed by default if the version of Python
is 2.7.9 or greater), it can be installed by running the following commands:
curl “https://bootstrap.pypa.io/get-pip.py” -o “get-pip.py”
python get-pip.py
After pip is installed, run the following command to install flask version 0.12.2:
pip install flask==0.12.2
After all the required software has been installed, in the same directory that the IBM
Spectrum Archive EE was extracted to, there is an RHEL7 directory that contains a file
called.ibmsa-rest-[version]-[buildlevel].x86_64.rpm. To install the restful service, run
the yum install command on this file, as shown in Example 4-10.
Dependencies Resolved
==================================================================================
===================================
Package Arch Version Repository
Size
==================================================================================
===================================
Installing:
ibmsa-rest x86_64 1.2.4.0-12441
/ibmsa-rest-1.2.4.0-12441.x86_64 51 k
Transaction Summary
==================================================================================
===================================
Install 1 Package
Total size: 51 k
Installed size: 51 k
86 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Downloading packages:
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
Installing : ibmsa-rest-1.2.4.0-12441.x86_64
1/1
##############################################################
# ibmsa-rest have been installed successfully.
# Please restart or reload httpd to enable REST server.
##############################################################
Verifying : ibmsa-rest-1.2.4.0-12441.x86_64
1/1
Installed:
ibmsa-rest.x86_64 0:1.2.4.0-12441
Complete!
At the bottom of a successful installation, it says the installation was successful and that a
restart of the httpd service is required to enable the rest server. To restart the service, run the
following command:
systemctl restart httpd
Note: If REST is already installed on cluster updating IBM Spectrum Archive EE will
automatically update the REST interface, however a manual restart of httpd is required
after starting IBM Spectrum Archive EE.
When this is all done, to quickly test that the rest service has been successfully installed, run
the following command:
curl -i -X GET ‘http://localhost:7100/ibmsa/v1’
Example 4-11 shows using a test curl command to see whether the installation was
successful.
The default port is on 7100 and the default protocol to use is http. If SSL is required,
uncomment SSLEngine, SSLCertificateFile, SSLCertifcateKeyFile and provide the direct
path to both the certificate file and the certificate key file in the following file:
/etc/httpd/conf.d/ibmsa-rest-httpd.conf
For an overview of IBM Spectrum Archive EE Rest API and commands, see 6.26, “IBM
Spectrum Archive REST API” on page 219.
The library replacement procedure requires the user to halt their environment to perform, due
to the relocation of all tape cartridges and possibly all tape drives to a newer library.
The pool relocation procedure requires a multi-library configuration so that a pool can be
relocated from one library to the other. While the pool is being relocated operations are not
allowed on that pool however the other pools are still available for operations.
In addition to these new procedures, IBM Spectrum Archive EE now appends a new state for
tape cartridges, appendable. This new state is determined by the pool settings and the state
of the tape to determine if the tape is candidate for migration. An appendable tape cartridge
within a pool allows data to be written to that tape. Those tapes can either be empty, or
partially filled with data.
Non-appendable tape cartridges are tapes that do not allow writes to the tape.
Non-appendable tape cartridges result from being completely full, erroneous tapes, or tapes
within pools that do not match their media-restriction or format type. With the introduction of
this new tape state, users have better control over the flow of their data from disk to tape and
from tape to tape. Refer to 6.11.5, “Data Migration” on page 180.
Note: The following library replacement procedures are supported with IBM Spectrum
Archive EE v1.2.6 and above.
88 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
4.6.1 Library replacement procedure
This procedure must be used if you are switching or upgrading your current tape library to
another tape library and want to continue using the tapes and tape drives and cluster
configuration within your IBM Spectrum Archive EE environment. For example, you might be
replacing your TS3500 tape library with a new TS4500 tape library.
2. If the tape drives from the old library will not be used in the new library, remove the tape
drives by using the eeadm drive unassign command. Refer to “IBM Spectrum Scale
commands” on page 323 for more information about the command.
3. Stop IBM Spectrum Archive EE by executing the eeadm cluster stop command.
4. Replace the physical old tape library with the new tape library.
5. Physically move all of the tape cartridges to the new library.
6. If you are moving the tape drives from the old tape library to the new library, physically
remove the tape drives from the old tape library and install them in the new library. The
drives need to be attached to the same node that they were attached to before.
7. The REPLACE_LIBRARY command associates the new library’s serial number with the
old library’s ID. Example 4-13 shows the output from running the ltfsee_config -m
REPLACE_LIBRARY command to associate the new library with the old library serial number.
90 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Note: In releases before IBM Spectrum Archive EE version 1.2.6, the serial number of a
library was used as the library ID. Beginning with version 1.2.6, the library ID can have a
unique value other than the serial number. A UUID is assigned as the library ID in libraries
that are configured with version 1.2.6 and subsequent releases.
With the separation of the library ID from the library serial number, IBM Spectrum Archive
EE can replace a library by changing the library serial number that is associated with the
library ID. The configuration of the cartridges, and pools, and drives in the old library is
transferred to the new library.
Note: Pool relocation will fail if any files are migrated to pools from IBM Spectrum Archive
EE v1.1. To determine if any pools have files from v1.1, run the following command:
ltfsee_count_v11_files -p <pool_name> -l <library_name>
Pool relocation only works in a multi-library IBM Spectrum Archive EE cluster. Therefore, if
you have a single-library environment and want to move a pool, you first need to have a
second IBM Spectrum Archive EE node added to the cluster and an identical media type
secondary tape library attached.
For the pool replacement procedure the following assumptions are made:
IBM Spectrum Archive EE is configured as multi-library configuration.
A pool is selected to move from one library (called the source library) to another library
(called the destination library).
All physical tape cartridges in the selected pool are moved to the destination library
manually.
1. Get the pool information that is being relocated by running the eeadm pool list
command.
2. Stop IBM Spectrum Archive EE.
Start the mmapplypolicy GPFS command to get a list of migrated files that are
in IBM Spectrum Archive V1.1 format.
Starting IBM Spectrum Archive EE to refresh the index cache of the libraries.
Started the IBM Spectrum Archive EE services for library lib1 with good status.
92 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Note: Pools with the same name appear in the source and destination libraries after
this command. (The pool name specified for relocation must not exist in the destination
library before this command). The mode attribute of the source pool is set to
relocation_source, and the mode attribute of the destination pool is set to
relocation_destination.
4. List pools that have been or are in process of being moved, as shown in Example 4-15.
5. Start IBM Spectrum Archive EE by using the eeadm cluster start command.
Note: Jobs related to the tapes in the pool with “source” or “destination” in the Mode
attribute will be rejected.
6. Make a list of tapes which belong to the pool for relocation with the eeadm tape list
command.
7. Move tapes within the selected pool to the IE slot to remove from the previous library using
the eeadm tape move <tape1[tape2 tape3 ...] -L ieslot -p <pool> -l <library>
command.
8. Stop IBM Spectrum Archive.
9. Activate the pool for relocation by running the ltfsee_config -m ACTIVATE_MOVE_POOL
command. Example 4-16 illustrates activating the relocated pool after all tapes have been
physically moved to the new tape library.
Starting IBM Spectrum Archive EE to refresh the index cache of the libraries.
Checking whether all the tapes in pool pool1 are properly moved from library
lib1 to lib2.
10.Verify that the pool has been activated, as shown in Example 4-17.
11.Start IBM Spectrum Archive EE by using the eeadm cluster start command.
94 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
12.Verify that the pool and tapes are valid, as shown in Example 4-18.
If upgrading a single node cluster or if the environment does not follow the Validity Checklist,
see Upgrading Operating System from RH7.x to RH8.x.
Similar node refresh techniques are described in 3.8, “High-level component upgrade steps”
on page 70. This information provides more references when planning or performing host
operating system upgrades.
Validity checklist
Use this check list to confirm whether the upgrade procedure is fully planned. All
requirements must met to perform the upgrade:
The cluster consists of more than one node per library.
All nodes in the cluster are upgraded to 1.3.1.2 or later, and have matching release
versions.
The cluster can handle requests as the nodes are removed.
Each node that is left in the cluster can reach at least one Control Path Drive on a library.
Caution: Ensure that IBM Spectrum Scale and IBM Spectrum Protect are upgraded to
ensure interoperability.
Note: These operating system upgrade steps refer only to the upgrade steps of IBM
Spectrum Archive EE. The operating system upgrade must be carefully planned to meet all
hardware and other software compatibility requirements. For operating system update
assistance, contact IBM.
96 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
5
Note: In the lab setup for this book, we used a Red Hat based Linux system. The screen
captures within this chapter are based on Version 1 Release 3 of the product. Although the
steps that you will perform are the same, you might see slightly different output responses
depending on your currently used version and release of the product.
Ensure that the following prerequisites are met before IBM Spectrum Archive EE is
configured. For more information, see 5.2, “Configuring IBM Spectrum Archive EE” on
page 105.
The Configuration worksheet is completed and available during the configuration process.
Tip: Table 5-1 on page 98, Table 5-2 on page 99, Table 5-3 on page 99, Table 5-4 on
page 100, Table 5-5 on page 100, and Table 5-6 on page 101 provide a set of sample
configuration worksheets. You can print and use these samples during your
configuration of IBM Spectrum Archive EE.
For more information, see 5.1.2, “Obtaining configuration information” on page 101 and follow
the steps to obtain the information that is required to complete your worksheet.
The information in the following tables is required to configure IBM Spectrum Archive EE.
Complete Table 5-4 on page 100, Table 5-5 on page 100, and Table 5-6 on page 101 with the
required information and refer to this information as necessary during the configuration
process, as described in 5.2, “Configuring IBM Spectrum Archive EE” on page 105.
Table 5-1, Table 5-2, and Table 5-3 show example configuration worksheets with the
parameters completed for the lab setup that was used to write this book.
File system name Mount point Need space Reserved for IBM
management? Spectrum Archive
(Yes or No) EE?
(Yes or No)
98 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Table 5-2 lists the logical tape library information.
IBM Spectrum Scale Installing IBM Tape drives assigned CP enabled tape
node name Spectrum Archive to this node drive
EE? (Serial number) (Serial number)
(Yes or No)
Figure 5-1 shows an example of a TS3500 GUI window that you use to display the starting
SCSI element address of a TS3500 logical library. You must record the decimal value
(starting address) to calculate the associated logical library serial number by using as shown
in Table 5-2 on page 99. You can open this window if you check for the details for a specific
logical library.
Figure 5-1 Obtain the starting SCSI element address of a TS3500 logical library
File system name Mount point Need space Reserved for IBM
management? Spectrum Archive
(Yes or No) EE?
(Yes or No)
100 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Table 5-6 shows an example a blank nodes worksheet.
IBM Spectrum Scale Installing IBM Tape drives assigned CP enabled tape
node name Spectrum Archive to this node drive
EE? (Serial number) (Serial number)
(Yes or No)
Keep the completed configuration worksheet available for reference during the configuration
process. Example 5-1 shows how to obtain the information for the worksheet.
Example 5-1 Obtain the IBM Spectrum Scale required information for the configuration worksheet
[root@ltfs97 ~]# mmstartup -a
Fri Apr 5 14:02:32 JST 2013: mmstartup: Starting GPFS ...
htohru9.ltd.sdl: The GPFS subsystem is already active.
102 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
-E yes Exact mtime mount option
-S relatime Suppress atime mount option
-K whenpossible Strict replica allocation option
--fastea yes Fast external attributes enabled?
--encryption no Encryption enabled?
--inode-limit 600000512 Maximum number of inodes
--log-replicas 0 Number of log replicas
--is4KAligned yes is4KAligned?
--rapid-repair yes rapidRepair enabled?
--write-cache-threshold 0 HAWC Threshold (max 65536)
--subblocks-per-full-block 512 Number of subblocks per full block
-P system Disk storage pools in file system
--file-audit-log no File Audit Logging enabled?
--maintenance-mode no Maintenance Mode enabled?
-d nsd208_209_1;nsd208_209_2;nsd208_209_3;nsd208_209_4 Disks in file
system
-A yes Automatic mount option
-o none Additional mount options
-T /ibm/glues Default mount point
--mount-priority 0 Mount priority
To use key-based login with OpenSSH, it is necessary to generate SSH key files and append
the public key file from each node (including the local node) to the authorized_keys file in the
~root/.ssh directory.
Important: You can verify whether the root user has a public key by locating the id_rsa
and id_rsa.pub files under the /root/.ssh/ directory. If these files do not exist, you
must generate them.
5.1.4 Preparing the IBM Spectrum Scale file system for IBM Spectrum
Archive EE
Complete this task to create and mount the IBM Spectrum Scale file system before IBM
Spectrum Archive EE is configured.
Before you make any system upgrades or major configuration changes to your GPFS or IBM
Spectrum Scale cluster, review your GPFS or IBM Spectrum Scale documentation and
consult IBM Spectrum Scale frequently asked question (FAQ) information that applies to your
version of IBM Spectrum Scale. For more information about the IBM Spectrum Scale FAQ,
see IBM Documentation.
Before you begin this procedure, ensure that the following prerequisites are met:
IBM Spectrum Scale is installed on each of the IBM Spectrum Archive EE nodes.
The IBM Spectrum Scale cluster is created and all of the IBM Spectrum Archive EE nodes
belong to the cluster.
IBM Spectrum Archive EE requires space for the file metadata, which is stored in the LTFS
metadata directory. The metadata directory can be stored in its own GPFS file system, or it
can share the GPFS file system that is being space-managed with IBM Spectrum Archive EE.
The file system that is used for the LTFS metadata directory must be created and mounted
before the IBM Spectrum Archive EE configuration is performed. The following requirements
apply to the GPFS file system that is used for the LTFS metadata directory:
The file system must be mounted and accessible from all of the IBM Spectrum Archive EE
nodes in the cluster.
The GPFS file system (or systems) that are space-managed with IBM Spectrum Archive
EE must be DMAPI enabled.
To create and mount the GPFS file system, complete the following steps:
1. Create a network shared disk (NSD), if necessary, by running the following command. It is
possible to share an NSD with another GPFS file system.
# mmcrnsd -F nsd.list -v no
<<nsd.list>>
%nsd: device=/dev/dm-3
nsd=nsd00
servers=ltfs01, ltfs02, ltfs03, ltfs04
usage=dataAndMetadata
2. Start the GPFS service (if it is not started already) by running the following command:
# mmstartup -a
3. Create the GPFS file system by running the following command. For more information
about the file system name and mount point, see 5.1.1, “Configuration worksheet tables”
on page 98.
# mmcrfs /dev/gpfs nsd00 -z yes -T /ibm/glues
104 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
In this example, /dev/gpfs is the file system name and /ibm/glues is the mount point. For
a separate file system that is used only for the LTFS metadata directory, you do not need
to use the -z option. Generally, if a GPFS file system is not intended to be IBM Spectrum
Archive EE managed, it should not be DMAPI-enabled; therefore, the -z option should not
be specified. The preferred configuration is to have one file system with DMAPI-enabled.
For more information about the inode size, see 7.5, “Preferred inode size for IBM
Spectrum Scale file systems” on page 239.
4. Mount the GPFS file system by running the following command:
# mmmount gpfs -a
For more information about the mmmount command, see the following resources:
General Parallel File System Version 4 Release 1.0.4 Advanced Administration Guide,
SC23-7032
IBM Spectrum Scale: Administration Guide, which is available at IBM Documentation.
Reminder: All of the command examples use the command without the full file path name
because we added the IBM Spectrum Archive EE directory (/opt/ibm/ltfsee/bin) to the
PATH variable.
The ltfsee_config command-line tool is shown in the following example and includes the
following options:
ltfsee_config -m <mode> [options]
-m
<mode> and [options] can be one of the following items:
– CLUSTER [-c]
Creates an IBM Spectrum Archive EE cluster environment and configures a
user-selected IBM Spectrum Scale (GPFS) file system to be managed by the IBM
Spectrum Archive or used for its metadata. The user must run this command one time
from one of the IBM Spectrum Archive nodes. Running the command a second time
modifies the file systems settings of the cluster.
Note: Even if you configure two control nodes per tape library, you still only run
ADD_CTRL_NODE once per tape library.
106 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Options:
– -a
Assigns the IP address of the Admin node name as the control node (MMM), or as the
local node. If -a is not used, the IP address of the Daemon node name is assigned.
– -c
Check and show the cluster or node configuration, without configuring or modifying it.
– -g
Assign the node to a node group that is selected or specified by user. If -g is not used,
the node is added to the default node group, which is named G0 if it did not exist before.
– -G
Specifies the node group to assign to the pool in the destination library during
translation between libraries, or the node group to be removed.
– -N
Remove a non-local node by specifying its node ID. If -N is not used, the local node is
removed.
– -f
Force node removal. If -f is not used, an attempt to remove a control node fails and the
configuration remains unchanged. When a control node is removed by using -f, other
nodes from the same library and the drives that are configured for those nodes are also
removed. To avoid removing multiple nodes, consider first setting another configured
non-control node from the same library as the control node (SET_CTRL_NODE).
Important: When the active control node is removed by use of -f option, the library
and the pool information stored in the internal database will be in-validated. If files
that are migrated only to that library are left in the system, recalls of those files will
no longer be possible.
– -b
Skips restarting the HSM daemon as a post process of the operation.
– -p
Specifies the name of the pool to be relocated to a different library.
– -P
Specifies the directory path storing SOBAR Support for System Migration result.
– -s
Specifies the name of the source library for a pool relocation procedure.
– -d
Specifies the destination library for a pool relocation procedure.
– -l
Specifies the library name in which to remove a node group with the -N option.
GPFSNative
GPFS NativeClient
Client GPFSNative
GPFS NativeClient
Client GPFS Native Client
GPFS Native Client GPFSNative
GPFS NativeClient
Client
NFS Client CIFS/SMB Client IBM Spectrum Scale Client FTP Client
Ethernet
EE Node Group 1
IBM EE Control
Spectrum Node 1
Scale
F F Legend
IBM Spectrum Archive Node
The steps in this section must be performed only on one node of an IBM Spectrum Archive
EE cluster environment. If you plan to have only one IBM Spectrum Archive EE node, this is a
so-called single-node cluster setup.
If you plan to set up a multi-node cluster environment for IBM Spectrum Archive EE, this
configuration mode must be performed once, and only on a node of your choice of your
cluster environment. All other nodes must be added. To do so, see 5.2.3, “Configuring a
multiple-node cluster” on page 112.
To configure a single-node cluster for IBM Spectrum Archive EE, complete the following
steps:
1. Log on to the operating system as a root user.
2. Start GPFS (if it is not already started) by running the following command:
# mmstartup -a
3. Mount the GPFS file system (if it is not already mounted) by running the following
command:
# mmmount all
4. Start the IBM Spectrum Archive EE configuration utility with the -m CLUSTER option by
running the following command and answering the prompted questions:
# ltfsee_config -m CLUSTER
108 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 5-2 shows the successful run of the ltfsee_config -m CLUSTER command during
the initial IBM Spectrum Archive EE configuration on the lab setup that was used for this
book. The example stops the process and will not proceed to the ADD_CTRL_NODE,
though you may answer [y] to continue to ADD_CTRL_NODE (step 5).
File system
1. /dev/gpfs Mount point(/ibm/gpfs) DMAPI(Yes)
q. Quit
** Select file systems to configure for IBM Spectrum Scale (GPFS) file system
for Space Management.
Input the corresponding numbers and press Enter
or press q followed by Enter to quit.
Press a followed by Enter to select all file systems.
Multiple file systems can be specified using comma or white space
delimiters.
File system
1. /dev/gpfs Mount point(/ibm/gpfs)
a. Select all file systems
q. Quit
Important: During the first run of the ltfsee_config -m CLUSTER command, if you see
the following error:
No file system is DMAPI enabled.
At least one file system has to be DMAPI enabled to use IBM Spectrum Archive
Enterprise Edition.
Enable DMAPI of more than one IBM Spectrum Scale (GPFS) file systems and try
again.
Ensure that DMAPI is turned on correctly, as described in 5.1.4, “Preparing the IBM
Spectrum Scale file system for IBM Spectrum Archive EE” on page 104. You can use
the following command sequence to enable DMAPI support for your GPFS file system
(here the GPFS file system name that is used is gpfs):
# mmumount gpfs
mmumount: Unmounting file systems ...
# mmchfs gpfs -z yes
# mmmount gpfs
mmmount: Mounting file systems ...
5. Run the IBM Spectrum Archive EE configuration utility by running the following command
and answering the prompted questions:
# ltfsee_config -m ADD_CTRL_NODE
Example 5-3 shows the successful run of the ltfsee_config -m ADD_CTRL_NODE
command during initial IBM Spectrum Archive EE configuration on the lab setup that was
used for this book.
110 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Configured IBM Spectrum Scale (GPFS) performance related settings.
If you are configuring multiple nodes for IBM Spectrum Archive EE, continue to 5.2.3,
“Configuring a multiple-node cluster” on page 112.
When configuring any multiple-node clusters, set a secondary node as a redundant control
node for availability features. The benefits of having redundancy are explained in 6.7, “IBM
Spectrum Archive EE automatic node failover” on page 153.
Figure 5-3 shows a multiple-node cluster configuration that is described in this section.
GPFSNative
GPFS NativeClient
Client GPFSNative
NativeClient
Client GPFS Native Client
GPFS Native Client GPFSNative
GPFS NativeClient
Client
NFS Client GPFS
CIFS/SMB Client IBM Spectrum Scale Client FTP Client
Ethernet
F F F F F F
SAN or
D D D D D D shared NSD access
Library
Pool 1 Pool 2
T1 T2 T3 T4 T5
Free Tapes
112 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Before configuring more nodes, ensure that all tasks that are described in 5.1, “Configuration
prerequisites” on page 98 are completed and that the first node of the cluster environment is
configured, as described in 5.2.2, “Configuring a single node cluster” on page 108.
To configure another node for a multi-node cluster setup for IBM Spectrum Archive EE,
complete the following steps:
1. Log on to the operating system as a root user.
2. Start GPFS (if it is not already started) by running the following command:
# mmstartup -a
3. Mount the GPFS file system on all nodes in the IBM Spectrum Scale cluster (if it is not
already mounted) by running the following command:
# mmmount all -a
4. Start the IBM Spectrum Archive EE configuration utility with the -m ADD_NODE option by
running the following command and answering the prompted questions:
# /opt/ibm/ltfsee/bin/ltfsee_config -m ADD_NODE
Important: This step must be performed on all nodes, except for the first node that was
configured in 5.2.2, “Configuring a single node cluster” on page 108.
Example 5-4 shows how to add a secondary node and set it as a redundant control node by
running ltfsee_config -m ADD_NODE. In step 5 of the command, after selecting which library
to add the node to, a prompt will appear asking to make the node a redundant control node.
Enter y to make the second node a redundant control node. Only two nodes per library can be
control nodes. If there are more than two nodes added to the cluster, enter n for each
additional node.
Input >y
The node ltfsml2(9.11.120.201) has been added as a control node for control node
redundancy
Added this node (ltfsml2, node id 2) to library ltfsee_lib.
114 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Do you want to use the saved configuration (y/n)?
Input > y
The LE+ component configuration is restored from a saved configuration.
Configured the LE+ component.
To summarize, you ran the following configuration options on EE Node 1 in 5.2.2, “Configuring
a single node cluster” on page 108:
ltfsee_config -m CLUSTER
ltfsee_config -m ADD_CTRL_NODE
For each additional IBM Spectrum Archive node in EE Node Group 1, run the ltfsee_config
-m ADD_NODE command. For example, in Figure 5-3 on page 112, you must run ltfsee_config
-m ADD_NODE on both EE Node 2 and EE Node 3.
The second tape library can be the same tape library model as the first tape library or can be
a different tape library model. These two tape libraries can be connected to a IBM Spectrum
Scale cluster in a single site or can be placed in metro distance (less than 300 km) locations
through IBM Spectrum Scale synchronous mirroring (stretched cluster).
For more information about synchronous mirroring by using IBM Spectrum Scale replication,
see IBM Documentation.
Important: Stretched cluster is available for distances shorter than 300 km. For longer
distances, the Active File Management (AFM) feature of IBM Spectrum Scale should be
used with IBM Spectrum Archive. The use of AFM is with two different IBM Spectrum Scale
clusters with one instance of IBM Spectrum Archive at each site. For more information
about IBM Spectrum Scale AFM support, see 2.2.5, “Active File Management” on page 30.
GPFSNative
GPFS Native Client GPFSNative
NativeClient
Client GPFS Native Client GPFSNative
Native Client
NFS ClientClient GPFS
CIFS/SMB Client IBM GPFS Native
Spectrum Client
Scale Client GPFS
FTP ClientClient
Ethernet
F F F F F F F F F F F F
SAN or
shared NSD access
D D D D D D D D D D D D
Library 1 Library 2
Pool 1 Pool 2 Pool A Pool B
T1 T2 T3 T4 T5 Ta Tb Tc Td Te
Figure 5-4 IBM Spectrum Archive multiple-node cluster configuration across two tape libraries
Before configuring more nodes, ensure that all tasks that are described in 5.1, “Configuration
prerequisites” on page 98 are completed and that the first node of the cluster environment is
configured, as described in 5.2.2, “Configuring a single node cluster” on page 108.
To configure the nodes at the other location for a multiple-node two-tape library cluster setup
for IBM Spectrum Archive EE, complete the following steps:
1. Run the IBM Spectrum Archive EE configuration utility by running the following command
and answering the prompted questions:
# /opt/ibm/ltfsee/bin/ltfsee_config -m ADD_CTRL_NODE
Using Figure 5-4 as an example, the ltfsee_config -m ADD_CTRL_NODE command is run
on EE Node 4.
2. Run the IBM Spectrum Archive EE configuration utility on all the remaining EE nodes at
the other location by running the following command and answering the prompted
questions:
# /opt/ibm/ltfsee/bin/ltfsee_config -m ADD_NODE
Using Figure 5-4 as an example, the ltfsee_config -m ADD_NODE command is run on
EE Node 5 and EE Node 6.
116 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Run eeadm cluster stop, to stop IBM Spectrum Archive EE and LE+. After IBM Spectrum
Archive EE has stopped, run ltfsee_config -m SET_CTRL_NODE to modify the configuration to
add a redundant control node.
Libraries
1:0000013FA0520411
2:0000013FA0520412
q.Quit
** Select 1 or 2 nodes for redundant control nodes from the following list.
They can be specified using comma or white space delimiters.
Nodes marked [x] are the current redundant configured nodes.
Nodes
1:[x]ltfsml1
2:[_]ltfsml2
q.Quit
Nodes
1:ltfsml1
2:ltfsml2
q.Quit
The node ebisu(9.11.120.198) has been set to be active for library ltfsee_lib
If HSM is already running on system, the output of this command should show the dsm
proccesses as shown in Example 5-7 . HSM is not running if no output is shown.
Information about how to start HSM can be referenced in 6.2.3, “Hierarchical Space
Management” on page 138.
2. Start the IBM Spectrum Archive EE program by running the following command:
/opt/ibm/ltfsee/bin/eeadm cluster start
Important: If the eeadm cluster start command does not return after several
minutes, it might be either because tapes are being unloaded or because the firewall is
running. The firewall service must be disabled on the IBM Spectrum Archive EE nodes.
For more information, see 4.3.2, “Installing, upgrading, or uninstalling IBM Spectrum
Archive EE” on page 78.
Example 5-7 shows all of the steps and the output when IBM Spectrum Archive EE was
started the first time. During the first start, you might discover a warning message, as shown
in the following example:
The authenticity of host 'localhost (::1)' can't be established.
RSA key fingerprint is 8f:56:95:fe:9c:eb:37:7f:95:b1:21:b9:45:d6:91:6b.
Are you sure you want to continue connecting (yes/no)?
This message is normal during the first start and you can easily continue by entering yes and
pressing Enter.
118 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 5-7 Start IBM Spectrum Archive EE the first time
[root@ltfsml1 ~]# ps -afe | grep dsm
root 14351 1 0 15:33 ? 00:00:01 /opt/tivoli/tsm/client/hsm/bin/dsmwatchd
nodetach
root 15131 30301 0 16:33 pts/0 00:00:00 grep --color=auto dsm
root 17135 1 0 15:33 ? 00:00:00 dsmrecalld
root 17160 17135 0 15:33 ? 00:00:00 dsmrecalld
root 17161 17135 0 15:33 ? 00:00:00 dsmrecalld
Now, IBM Spectrum Archive EE is started and ready for basic usage. For further handling,
managing, and operations of IBM Spectrum Archive EE (such as creating pools, adding and
formatting tapes, and setting up migration policies), see Chapter 6, “Managing daily
operations of IBM Spectrum Archive Enterprise Edition” on page 129.
5.3.1 Configuring IBM Spectrum Archive EE with IBM Spectrum Scale AFM
This section walks through how to set up IBM Spectrum Archive EE and IBM Spectrum Scale
AFM to create either a Centralized Archive Repository, or an Asynchronous Archive
Replication solution. The steps shown in this section assume that the user has already
installed and configured IBM Spectrum Archive EE.
If IBM Spectrum Archive EE has not been previously installed and configured, set up AFM
first and then follow the instructions in Chapter 4, “Installing IBM Spectrum Archive Enterprise
Edition” on page 73 to install Spectrum Archive EE and then in Chapter 5, “Configuring IBM
Spectrum Archive Enterprise Edition” on page 97. If performed in that order, you can skip this
section. See 7.10.3, “IBM Spectrum Archive EE migration policy with AFM” on page 246 for
information about creating migration policy on cache nodes.
Important: Starting with IBM Spectrum Archive EE V1.2.3.0, IBM Spectrum Scale AFM is
supported. This support is limited to only one cache mode, independent writer (IW).
For more information about configuring IBM Spectrum Scale AFM, see the AFM
documentation at IBM Documentation.
Steps 1 - 5 demonstrate how to set up a IBM Spectrum Scale AFM home cluster and start
IBM Spectrum Archive EE. Steps 6 - 9 show how to set up the IW caches for IBM Spectrum
Scale AFM cache clusters:
1. If IBM Spectrum Scale is not already active and GPFS is not already mounted, start IBM
Spectrum Scale and wait until the cluster becomes active. Then, mount the file system if it
is not set to mount automatically using the commands in Example 5-8.
Example 5-8 Starting and mounting IBM Spectrum Scale and GPFS file system
[root@ltfseehomesrv ~]# mmstartup -a
Tue Mar 21 14:37:57 MST 2017: mmstartup: Starting GPFS ...
[root@ltfseehomesrv ~]# mmgetstate -a
Note: Step 2 assumes that the user has already created their file set and linked it to the
GPFS file system. The following examples use IWhome as the home file set.
120 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2. After IBM Spectrum Scale is active and the GPFS file system is mounted, edit the NFS
exports file (/etc/exports) to include the new file set. It is important that the
no_root_squash, sync, and rw arguments are used. Example 5-9 shows example content
of the exports file for file set IWhome.
Note: The fsid in the exports file needs to be a unique number different than any other
export clause within the exports file.
3. After the exports file has been modified to include the file set, start the NFS service.
Example 5-10 shows an example of starting and checking the NFS service.
4. After NFS has properly started the final step to configure IBM Spectrum Scale AFM at the
home cluster is to enable the exported path. Run mmafmconfig enable <path-to-fileset>
to enable the exported file set. Example 5-11 shows the execution of the mmafmconfig
command with the IWhome file set.
5. After the file set has been enabled for AFM proceed by starting up IBM Spectrum Archive
if it has not been started previously by running eeadm cluster start.
7. After all the gateway nodes have been set, start IBM Spectrum Scale and mount the file
system if it is not done automatically, as shown in Example 5-13:
a. mmmstartup -a
b. mmgetstate -a
c. mmmount all -a (optional, only if the GPFS file system is not mounted automatically)
Example 5-13 Starting and mounting IBM Spectrum Scale and GPFS file system
[root@ltfseecachesrv ~]# mmstartup -a
Tue Mar 21 14:37:57 MST 2017: mmstartup: Starting GPFS ...
[root@ltfseecachesrv ~]# mmgetstate -a
8. After IBM Spectrum Scale has been started and the GPFS file system is mounted, then
create the cache fileset by using mmcrfileset with the afmTarget, afmMode, and
inode-space parameters. Example 5-14 shows the execution of mmcrfileset to create a
cache fileset.
Example 5-14 Creating a cache fileset that targets the home fileset
[root@ltfseecachesrv ~]# mmcrfileset gpfs iwcache -p afmmode=iw -p
afmtarget=ltfseehomesrv:/ibm/glues/IWhome --inode-space=new
Fileset iwcache created with id 1 root inode 4194307.
9. After the fileset is created, it can be linked to a directory in the GPFS file system by
running the mmlinkfileset <device> <fileset> -J <gpfs file system/fileset name>
command. Example 5-15 shows output of running mmlinkfileset.
Example 5-15 Linking the GPFS fileset to a directory on the GPFS file system
[root@ltfseecachesrv glues]# mmlinkfileset gpfs iwcache -J /ibm/glues/iwcache
Fileset iwcache linked at /ibm/glues/iwcache
122 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Steps 6 - 9 need to be run on each cache cluster that will be linked to the home cluster. After
completing these steps, IBM Spectrum Scale AFM and IBM Spectrum Archive EE are set up
on the home cluster and IBM Spectrum Scale AFM is set up on each cache cluster. The
system is ready to perform centralized archiving and caching.
Steps 1 - 5 demonstrate how to set up a IBM Spectrum Scale AFM home cluster and start
IBM Spectrum Archive EE. Steps 6 - 11 demonstrate how to set up the cache clusters, and
steps 12 - 15 demonstrate how to reconfigure Spectrum Archive EE’s configuration to work
with IBM Spectrum Scale AFM.
1. If IBM Spectrum Archive is not already active and GPFS is not already mounted, start the
file system and wait until the file system becomes active. Then, mount the file system if it is
not set to mount automatically using the commands in Example 5-16.
Example 5-16 Starting and mounting IBM Spectrum Scale and GPFS file system
[root@ltfseehomesrv ~]# mmstartup -a
Tue Mar 21 14:37:57 MST 2017: mmstartup: Starting GPFS ...
[root@ltfseehomesrv ~]# mmgetstate -a
2. After IBM Spectrum Scale is active and the GPFS file system is mounted, edit the NFS
exports file (/etc/exports) to include the new fileset. It is important that the
no_root_squash, sync, and rw arguments are used. Example 5-17 shows example content
of the exports file for fileset IWhome.
Note: The fsid in the exports file needs to be a unique number different than any other
export clause within the exports file.
3. After the exports file has been modified to include the fileset, start the NFS service.
Example 5-18 shows an example of starting and checking the NFS service.
4. After NFS has properly started, the final step to configure IBM Spectrum Scale AFM at the
home cluster is to enable the exported path. Run mmafmconfig enable <path-to-fileset>
to enable the exported fileset. Example 5-19 shows the execution of the mmafmconfig
command with the IWhome fileset.
124 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
5. After the fileset has been enabled for AFM, start Spectrum Archive if it has not been
started previously by running eeadm cluster start.
After the home cluster is set up and an NFS export directory is enabled for IBM Spectrum
Scale AFM, steps 6 - 11 demonstrate how to set up a Spectrum Scale AFM IW cache
fileset at a cache cluster and connect the cache’s fileset with the home’s fileset. Steps 12 -
15 show how to modify IBM Spectrum Archive EE’s configuration to allow cache file sets.
6. If IBM Spectrum Archive EE is active, properly shut it down by using the commands in
Example 5-20.
7. If IBM Spectrum Scale is active, properly shut it down by using the commands in
Example 5-21.
8. With IBM Spectrum Archive EE and IBM Spectrum Scale both shut down, set the gateway
nodes if they have not been set when IBM Spectrum Scale was configured by using the
command in Example 5-22.
9. Properly start IBM Spectrum Scale by using the commands in Example 5-23.
11.Link the fileset to a directory on the node’s GPFS file system by using the command in
Example 5-25.
IBM Spectrum Scale AFM is now configured and has a working home and IW cache
clusters.
126 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
12.With IBM Spectrum Archive EE still shut down, obtain the metadata and HSM file systems
IBM Spectrum Archive EE by using the command in Example 5-26.
13.Modify IBM Spectrum Archive EE’s configuration by using the command that is shown in
Example 5-27 with the same file systems that were recorded from step 7.
Example 5-27 Modify IBM Spectrum Archive EE configuration for IBM Spectrum Scale AFM
[root@ltfseecachesrv ~]# ltfsee_config -m UPDATE_FS_INFO
The EE configuration script is starting: /opt/ibm/ltfsee/bin/ltfsee_config -m
UPDATE_FS_INFO
UPDATE_FS_INFO mode starts .
## Step-1. Check to see if the cluster is already created ##
The cluster is already created and the configuration file
ltfsee_config.filesystem exists.
Successfully validated the prerequisites.
** Select file systems to configure for the IBM Spectrum Scale (GPFS) file
system for Space Management.
Input the corresponding numbers and press Enter
or press 'q' followed by Enter to quit.
Press a followed by Enter to select all file systems.
Multiple file systems can be specified using comma or white space delimiters.
File system
1. /dev/flash Mount point(/flash)
2. /dev/gpfs Mount point(/ibm/gpfs)
a. Select all file systems
q. Quit
Input number > a
## Step-2. Add selected file systems to Space Management ##
Added the selected file systems to the space management.
## Step-3. Store the file systems configuration and dispatch it to all nodes ##
15.At the start of IBM Spectrum Archive EE, the AFMSKIPUNCACHEDFILES flag inside the
/opt/tivoli/tsm/client/ba/bin/dsm.sys file should be set to yes. It can be checked by
using the command in Example 5-29. If it has not been properly set, modify the file so that
the AFMSKIPUNCACHEDFILES flag is set to yes.
After successfully completing these steps, IBM Spectrum Archive EE and IBM Spectrum
Scale AFM are set up at both the home and cache cluster. They can now be used as an
Asynchronous Archive Replication solution.
128 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 129
6.26, “IBM Spectrum Archive REST API” on page 219
6.27, “File system migration” on page 231
Note: The steps that you perform are the same as we describe in this Redbooks
publication. You might see slightly different output responses in your environment,
depending on your version and release of the product.
130 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.1 Overview
The following terms specific to IBM Spectrum Archive EE operations are used in this chapter:
Migration The movement of files from the IBM Spectrum Scale file system on disk
to IBM Linear Tape File System tape cartridges, which leaves behind a
stub file.
Premigration The movement of files from GPFS file systems on disk to LTFS tape
cartridges without replacing them with stub files on the GPFS file system.
Identical copies of the files are on the GPFS file system and in LTFS
storage.
Recall The movement of migrated files from tape cartridges back to the
originating GPFS file system on disk, which is the reverse of migration.
Reconciliation The process of synchronizing a GPFS file system with the contents of an
LTFS tape cartridge and removing old and obsolete objects from the tape
cartridge. You must run reconciliation when a GPFS file is deleted,
moved, or renamed.
Reclamation The process of defragmenting a tape cartridge. The space on a tape
cartridge that is occupied by deleted files is not reused during normal
LTFS operations. New data is always written after the last index on tape.
The process of reclamation is similar to the same named process in IBM
Spectrum Protect from the IBM Spectrum Archive family. All active files
are consolidated onto a second tape cartridge, which improves overall
tape usage.
Library rescan The process of triggering IBM Spectrum Archive EE to retrieve
information about physical resources from the tape library. This process
is scheduled to occur automatically at regular intervals, but can be run
manually.
Tape validate The process of validating the current condition of a tape by loading it to
the tape drive and updating the tape state.
Tape replace The process of moving the contents of a tape that previously suffered
some error to another tape in the same pool.
Import The addition of an LTFS tape cartridge to IBM Spectrum Archive EE.
Export The removal of an LTFS tape cartridge from IBM Spectrum Archive EE.
Data migration The new method of technology migration of tape drives and cartridges.
There are two ways to perform data migration:
- within a pool
- pool to pool
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 131
The following eeadm command options are available. All options, except listing resource
types, can be run only with root user permissions:
eeadm cluster start
Use this command to start the process of the IBM Spectrum Archive EE system on all
configured servers, or on a specific library.
Important: If the eeadm cluster start command does not return after several minutes, it
might be because the firewall is running or tapes are being unmounted from the drives.
The firewall service must be disabled on the IBM Spectrum Archive EE nodes. For
more information, see 4.3.2, “Installing, upgrading, or uninstalling IBM Spectrum
Archive EE” on page 78.
132 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
– premigrated: The data is both on disk and tapes
– migrated: The data is on tapes while its stub file remains on disk.
eeadm library list
Use this command to list all the managed tape libraries.
eeadm library rescan
Use this command to force the tape library to check and report its physical resources, and
update the resource information kept in IBM Spectrum Archive EE.
eeadm library show
Use this command to display the configuration attributes of a tape library
eeadm node down
Use this command to disable one of the IBM Spectrum Archive EE servers temporarily for
maintenance. The disabled node does not participate in the system.
eeadm node list
Use this command to list the configuration and status of all the configured nodes.
eeadm node show
Use this command to display the configuration attributes of the node.
eeadm node up
Use this command to enable one of the IBM Spectrum Archive EE servers. The enabled
node can be used as a part of the IBM Spectrum Archive EE system.
eeadm nodegroup list
Use this command to list all the configured node groups.
eeadm pool create
Use this command to create a tape pool.
eeadm pool delete
Use this command to delete a tape pool to which no tapes are assigned.
eeadm pool list
Use this command to list all the configured tape pools
eeadm pool set
Use this command to change the configuration attributes of a pool.
eeadm pool show
Use this command to display the configuration attributes of the tape pool.
eeadm tape list
Use this command to list the configuration and status of all the tapes.
eeadm tape set
Use this command to change the configuration attribute of a tape.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 133
eeadm tape assign
Use this command to format the tapes and assign it to the tape pool. The command fails if
it contains any file object unless the -f option is used. Use the eeadm tape import
command to make the tape that contains files a new member of the pool.
eeadm tape export
Use this command to export the tape permanently from the IBM Spectrum Archive EE
system and purges the GPFS files referring to the tape. The command internally runs the
reconciliation process and identifies the active GPFS files in migrated or premigrated
states. If a file refers to the tape to be exported and if the tape contains the last replica of
the file, it deletes the GPFS file. After the successful completion of the command, the tape
is unassigned from the tape pool, and the tape state becomes exported. Prior to running
this command, all GPFS file systems must be mounted so that the command is able to
check for the existence of the files on disk.
eeadm tape import
Use this command to import tapes created and managed by other system to the IBM
Spectrum Archive EE system, and makes the files on tape accessible from the GPFS
namespace. The command will create the stub files and leave the files in the migrated
state without transferring the file data back to disk immediately.
eeadm tape move
Use this command to move the tape physically within the tape library. The command can
move the tape to its home slot either from the tape drive or from the I/E slot (or I/O station)
of the tape library. It can move the tape to I/E slot if the tape is:
– in the offline state
– the tape belongs to a tape pool that is undergoing pool relocation
– tape is currently not assigned to a pool
eeadm tape offline
Use this command to set the tape to the offline state to prepare for moving the tape
temporarily out of the tape library, until access to the data is required. The tape needs to
come back to the original IBM Spectrum Archive EE system by using the eeadm tape
online command.
eeadm tape online
Use this command to make the offline tape accessible from the IBM Spectrum Archive EE
system.
eeadm tape unassign
Use this command to unassign the member tapes from the tape pool.
eeadm tape datamigrate
Use this command to move the active contents of specified tapes to a different tape pool
and updates the stub files on disk to point to the new data location. After the successful
completion of the command, the tape is automatically unassigned from the source tape
pool. The datamigrate command can be used to move the data on older technology tapes
in the source tape pool to newer technology tapes in the destination tape pool.
eeadm tape reclaim
Use this command to reclaim the unreferenced space of the specified tapes. It moves the
active contents on the tapes to different tapes, then recycles the specified tapes. If the
--unassign option is specified, the tape is automatically unassigned from the tape pool
after the successful completion of the.eeadm tape reconcile command.
134 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Use this command to compare the contents of the tape with the files on the GPFS file
systems, and reconciles the differences between them.
eeadm tape replace
Use this command to move the contents of a tape that previously suffered some error to
another tape in the same pool. This command is used on tapes in the require_replace or
need_replace state. After the successful completion of the replacement process, the tape
that had the error is automatically unassigned from the tape pool.
eeadm tape validate
Use this command to validate the current condition of a tape by loading it to the tape drive,
and update the tape state. The tape must be a member of the tape pool and online.
eeadm task cancel
Use this command to cancel the active task. The command supports the cancellation of
reclaim and datamigrate tasks only.
eeadm task clearhistory
Use this command to delete the records of completed tasks to free up disk space.
eeadm task list
Use this command to list the active or completed tasks.
eeadm task show
Use this command to show the detailed information of the specified task.
eeadm migrate
Use this command to move the file data to the tape pools to free up disk space, and sets
the file state to migrated.
eeadm premigrate
Use this command to copy the file data to the tape pools, and sets the file state to
premigrated.
eeadm recall
Use this command to recall the file data back from the tape and places the file in the
premigrated state, or optionally in the resident state.
eeadm save
Use this command to save the name of empty files, empty directories, and symbolic links
on the tape pools.
In addition, the CLI is used by the IBM Spectrum Scale mmapplypolicy command to trigger
migrations or premigrations. When this action occurs, the mmapplypolicy command calls IBM
Spectrum Archive EE when an IBM Spectrum Scale scan occurs, and passes the file name of
the file that contains the scan results and the name of the target tape cartridge pool.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 135
The eeadm command uses the following two syntax:
eeadm <resource_type> <action> [options], and eeadm <subcommand> [options]
Reminder: All of the command examples use the command without the full file path name
because we added the IBM Spectrum Archive EE directory (/opt/ibm/ltfsee/bin) to the
PATH variable.
The components should be checked in the order that is shown here because a stable, active
GPFS file system is a prerequisite for starting IBM Spectrum Archive EE.
This section describes how to obtain GPFS daemon state information by running the GPFS
command mmgetstate. For more information about the other GPFS commands, see the
following publications:
General Parallel File System Version 4 Release 1.0.4 Advanced Administration Guide,
SC23-7032
IBM Spectrum Scale: Administration Guide, which is available at IBM Documentation.
The node on which the mmgetstate command is run must have the GPFS mounted. The node
must also run remote shell commands on any other node in the GPFS/IBM Spectrum Scale
cluster without the use of a password and without producing any extraneous messages.
Example 6-1 shows how to get status about the GPFS/IBM Spectrum Scale daemon on one
or more nodes.
The -a argument shows the state of the GPFS/IBM Spectrum Scale daemon on all nodes in
the cluster.
136 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Permissions: Retrieving the status for GPFS/IBM Spectrum Scale requires root user
permissions.
The following GPFS/IBM Spectrum Scale states are recognized and shown by this command:
Active: GPFS/IBM Spectrum Scale is ready for operations.
Arbitrating: A node is trying to form a quorum with the other available nodes.
Down: GPFS/IBM Spectrum Scale daemon is not running on the node or is recovering
from an internal error.
Unknown: Unknown value. The node cannot be reached or some other error occurred.
If the GPFS/IBM Spectrum Scale state is not active, attempt to start GPFS/IBM Spectrum
Scale and check its status, as shown in Example 6-2.
If the status is active, also check the GPFS/IBM Spectrum Scale mount status by running the
command that is shown in Example 6-3.
Example 6-4 Check the IBM Spectrum Archive LE component status (running)
[root@ltfs97 ~]# df -m
Filesystem 1M-blocks Used Available Use% Mounted on
/dev/mapper/VolGroup-lv_root
33805 5081 27007 16% /
tmpfs 1963 0 1963 0% /dev/shm
/dev/vda1 485 36 424 8% /boot
/dev/gpfs 153600 8116 145484 6% /ibm/glues
ltfs:/dev/sg2 2147483648 0 2147483648 0% /ltfs
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 137
To start IBM Spectrum Archive LE run the eeadm cluster start command. If errors occur
during the start of the IBM Spectrum Archive EE system, run the eeadm node list command
to display which component failed to start. For more information about the updated
eeadm node list command, see 6.7, “IBM Spectrum Archive EE automatic node failover” on
page 153.
The dsmmigfs command also provides the status of HSM, as shown by the output in
Example 6-6.
138 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Min Partial Rec Size:0
Min Stream File Size:0
Min Mig File Size: 0
Inline Copy Mode: MIG
Preferred Node: yurakucho Node ID: 1
Owner Node: yurakucho Node ID: 1
You can also ensure that the GPFS file system (named gpfs in this example) is managed by
HSM by running the command that is shown in Example 6-7.
To manage a file system with IBM Spectrum Archive EE, it must be data management
application programming interface (DMAPI) enabled. A file system is managed by IBM
Spectrum Archive EE by running the ltfsee_config command, which is described in 5.2,
“Configuring IBM Spectrum Archive EE” on page 105.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 139
If the HSM watch daemon (dsmwatchd) is not running, Example 6-8 shows you how to start it.
If the HSM recall daemons (dsmrecalld) are not running, Example 6-9 shows you how to start
them.
If failover operations within the IBM Spectrum Scale cluster are wanted on the node, run the
dsmmigfs enablefailover command after you run the dsmmigfs start command.
Permissions: Retrieving the status for the MMM service does not require root user
permissions.
If the MMM service is running correctly, you see a message similar to the message shown in
Example 6-10.
140 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
If the MMM service is not running correctly, you may see a message that is similar to the
message shown in Example 6-11.
Spectrum Archive EE service (MMM) for library libb fails to start or is not
running on lib_ltfsml1 Node ID:1
Problem Detected:
Node ID Error Modules
1 LE; MMM;
In Example 6-11, IBM Spectrum Archive EE failed to start MMM because it was unable to
mount LE. In this example, MMM failure to mount LE was caused by the server having no
control path drives connected; therefore, the work around is to assign a control path drive to
the server and have the monitor daemon automatically mount LE.
To monitor the process of IBM Spectrum Archive EEs start, run the eeadm node list
command. If IBM Spectrum Archive EE is taking too long to recover, stop and start the
process by using the eeadm cluster stop/start command. Example 6-12 shows the process
of IBM Spectrum Archive EE recovering after a control path drive was connected.
Spectrum Archive EE service (MMM) for library libb fails to start or is not
running on lib_ltfsml1 Node ID:1
Problem Detected:
Node ID Error Modules
1 LE(Starting); MMM;
Important: If the eeadm cluster start command does not return after several minutes, it
might be because the firewall is running or tapes are being unmounted from the drives.
The firewall service must be disabled on the IBM Spectrum Archive EE nodes. For more
information, see 4.3.2, “Installing, upgrading, or uninstalling IBM Spectrum Archive EE” on
page 78.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 141
6.3 Upgrading components
The following sections describe the process that is used to upgrade IBM Spectrum Scale and
other components of IBM Spectrum Archive EE.
Before any system upgrades or major configuration changes are made to your IBM Spectrum
Scale cluster, review your IBM Spectrum Scale documentation and consult the IBM Spectrum
Scale frequently asked question (FAQ) information that applies to your version of IBM
Spectrum Scale.
For more information about the IBM Spectrum Scale FAQ, see IBM Documentation and select
the IBM Spectrum Scale release under the Cluster product libraries topic in the navigation
pane that applies to your installation.
2. Run the pidof mmm command on all EE Control Nodes until all MMM processes have been
terminated.
3. Run the pidof ltfs command on all EE nodes until all ltfs processes are stopped.
4. Disable DSM failover by running the command that is shown in Example 6-14.
5. Stop the IBM Spectrum Protect for Space Management HSM by running the command
that is shown in Example 6-15.
142 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Command Line Space Management Client Interface
Client Version 8, Release 1, Level 11.0
Client date/time: 03/06/21 05:35:38
(c) Copyright by IBM Corporation and other(s) 1990, 2020. All Rights Reserved.
If the mmumount all command results show that processes are still being used (as shown
in Example 6-18), you must wait for them to finish and then run the mmumount all
command again.
Example 6-18 Processes running that prevent the unmounting of the GPFS file system
[root@ltfs97 ~]# mmumount all
Tue Apr 16 23:46:12 JST 2013: mmumount: Unmounting file systems ...
umount: /ibm/glues: device is busy.
(In some cases useful info about processes that use
the device is found by lsof(8) or fuser(1))
umount: /ibm/glues: device is busy.
(In some cases useful info about processes that use
the device is found by lsof(8) or fuser(1))
8. Shut down GPFS by running the command that is shown in Example 6-19.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 143
9. Download the IBM Spectrum Scale update from IBM Fix Central. Extract the IBM
Spectrum Scale .rpm files and install the updated .rpm files by running the command that
is shown in Example 6-20.
10.Rebuild and install the IBM Spectrum Scale portability layer by running the command that
is shown in Example 6-21.
12.Mount the GPFS file system by running the command that is shown in Example 6-23.
13.Start the watch daemon by running the command that is shown in Example 6-24.
144 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
15.Enable failover by running the command that is shown in Example 6-26.
16.Start IBM Spectrum Archive EE by running the command that is shown in Example 6-27.
Optionally, you can check the status of each component when it is started, as described in
6.2, “Status information” on page 136.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 145
6.4 Starting and stopping IBM Spectrum Archive EE
This section describes how to start and stop IBM Spectrum Archive EE.
For example, to start IBM Spectrum Archive EE, run the command that is shown in
Example 6-28.
Important: If the eeadm cluster start command does not return after several minutes, it
might be because the firewall is running or unmounting tapes from drives. The firewall
service must be disabled on the IBM Spectrum Archive EE nodes. For more information,
see 4.3.2, “Installing, upgrading, or uninstalling IBM Spectrum Archive EE” on page 78.
You can confirm that IBM Spectrum Archive EE is running by referring to the steps in
Example 6-10 on page 140 or by running the command in Example 6-29.
Example 6-29 Check the status of all available IBM Spectrum Archive EE nodes
[root@ltfsml1 ~]# eeadm node list
Node ID State Node IP Drives Ctrl Node Library Node Group Host
Name
1 available 9.11.244.46 3 yes(active) libb G0
lib_ltfsml1
146 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.4.2 Stopping IBM Spectrum Archive EE
The eeadm cluster stop command stops the IBM Spectrum Archive EE system on all EE
Control Nodes.
For example, to start IBM Spectrum Archive EE, run the command that is shown in
Example 6-30.
In some cases, you might see the GLESM658I informational message if there are active
tasks on the task queue in IBM Spectrum Archive EE:
There are still tasks in progress.
To terminate IBM Spectrum Archive EE for this library,
run the "eeadm cluster stop" command with the "-f" or "--force" option.
If you are sure that you want to stop IBM Spectrum Archive EE, run the eeadm cluster stop
command with the -f option, which forcefully stops any running IBM Spectrum Archive EE
tasks abruptly. Note that this option will stop all tasks that are currently running on EE, and
will need to be manually re-run if necessary once the cluster has restarted.
Within each status group, the entries will be sorted by Priority (H, M, L order).
– H will be any transparent or selective recall commands
– M will be any premigration, migration, or save commands
– L will be everything else (general commands)
Within Running status, within the same Priority, the entries are sorted by Started Time (oldest
at the top).
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 147
Example 6-31 shows active tasks.
The other common use will be to get the list of completed tasks which returns the prior task
IDs, result, and date/time information, sorted by completed time (oldest at the top). For
administrators, this allows a quick view into task IDs, the history of what task has been
executed recently and their results. For more information about the specified task, see
“eeadm task show” on page 148.
For more information about the task status, see “User Task Reporting” on page 7.
148 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Accepted Time: Mon Jan 7 11:58:34 2019 (-0700)
Started Time: Mon Jan 7 11:58:34 2019 (-0700)
Completed Time: -
In-use Resources: 1068045923(D00369L5):pool1:G0:libb
Workload: 100 files. 1 replicas.
7545870371 bytes to copy. 1 copy tasklets on pool1@libb.
Progress: -
0/1 copy tasklets completed on pool1@libb.
Result Summary: -
Messages:
2019-01-07 11:58:34.231005 GLESM896I: Starting the stage 1 of 3 for migration
task 18830 (qualifying the state of migration candidate files).
2019-01-07 11:58:37.133670 GLESM897I: Starting the stage 2 of 3 for migration
task 18830 (copying the files to 1 pools).
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 149
Result Summary: 100 succeeded, 0 failed, 0 duplicate, 0 duplicate wrong
pool, 0 not found, 0 too small, 0 too early.
(GLESM899I) All files have been successfully copied on
pool1/libb.
Messages:
2019-01-07 11:58:34.231005 GLESM896I: Starting the stage 1 of 3 for migration
task 18830 (qualifying the state of migration candidate files).
2019-01-07 11:58:37.133670 GLESM897I: Starting the stage 2 of 3 for migration
task 18830 (copying the files to 1 pools).
2019-01-07 11:59:30.564568 GLESM898I: Starting the stage 3 of 3 for migration
task 18830 (changing the state of files on disk).
2019-01-07 11:59:34.123713 GLESL038I: Migration result: 100 succeeded, 0 failed,
0 duplicate, 0 duplicate wrong pool, 0 not found, 0 too small to qualify for
migration, 0 too early for migration.
The other common use will be to show the file results of each individual file from any
premigration, migration or recall tasks. With this information, the administrator can determine
which files were successful and which files failed including the error code (reason) and
date/time the error occurred. The administrator can quickly determine which files failed, and
take corrective actions including resubmission for those failed files. Example 6-35 shows the
completed task results.
150 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Success - - -
/ibm/gpfs/prod/LTFS_EE_FILE_h7SaXit1Of9vrUo_3yYT.bin
Success - - -
/ibm/gpfs/prod/LTFS_EE_FILE_RPrsQ0xxKAu3nJ9_xResu.bin
Success - - -
/ibm/gpfs/prod/LTFS_EE_FILE_CLS7aXD9YBwUNHhhfLlFSaVf4q7eMBtwHYVnMpcAWR6XwnPYL_rsQ.
bin
Success - - -
/ibm/gpfs/prod/LTFS_EE_FILE_XEvHOLABXWx4CZY7cmwnvyT9W5i5uu_bUvNC.bin
Success - - -
/ibm/gpfs/prod/LTFS_EE_FILE_3_wjhe.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_7QVSfmURbFlkQZJAYNvlPx82frnUelfyKSH0c7ZqJNsl_swA.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_8hB1_B.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_nl9T7Y4Z_1.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_W2E77x4f3CICypMbLewnUzQq91hDojdQVJHymiXZuHMJKPY_X.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_sOyPUWwKaMu3Y_VzS.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_tc5xwElJ1SM_x.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_yl_73YEI.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_UR65_nyJ.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_DXhfSFK8N2TrN7bhr0tNfNARwT3K1tZbp5SmBb8RbK_d.bin
Fail GLESC012E 2019/01/07T12:06:49 1
/ibm/gpfs/prod/LTFS_EE_FILE_ZRthniqdYS70yoblcUKc9uz9NECTtLnC8lNODxsQhj_QEWas.bin
For more information about the task status, see “User Task Reporting” on page 7.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 151
6.6 IBM Spectrum Archive EE database backup
IBM Spectrum Archive EE uses databases to store important system states and
configurations for healthy operations. Starting with IBM Spectrum Archive EE v1.2.4.0,
database backup is performed automatically and stores as many as three backups. The
backup files can be found under the /var/opt/ibm/ltfsee/local/dbbackup directory.
The backups are performed when IBM Spectrum Archive EE is running, part of the monitor
daemon checks periodically if the original database files have been modified by the above
operations. If the thread detects that there has been a modification, then it will go into a grace
period. If no further modifications are performed during a set amount of time during this grace
period, a backup will then occur.
This technique is intended to prevent repeating backups in a short amount of time, for
example when multiple tapes are being added to a pool when users perform eeadm tape
assign on multiple tapes. Instead, it performs a backup after the command finished executing
and no further changes are made in a set time limit.
These backups are crucial in rebuilding an IBM Spectrum Archive EE cluster if a server gets
corrupted or the GPFS file system needs to be rebuilt. After reinstalling the IBM Spectrum
Archive EE onto the server, replace the .global and .lib database files under the <path to
GPFS filesysytem>.ltfsee/config directory with the database files backed up from the
/var/opt/ibm/ltfsee/local/dbbackup/ directory.
152 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.7 IBM Spectrum Archive EE automatic node failover
This section describes IBM Spectrum Archive EE automatic failover features, and the new
LTFSEE monitoring daemon and the updated commands to display any nodes that are having
issues.
Components, such as MMM, IBM Spectrum Archive LE, and remote monitoring daemons,
have automatic recovery features. If one of those three components is hung, or has crashed,
the monitor daemon performs a recovery to restart it. In an environment where a redundant
control node is available and MMM is no longer responding or alive, an attempt to restart the
MMM service on the current node is made and if it fails a failover takes place and the
redundant control node becomes the new active control node.
If there is only one control node available in the cluster, then an in-place failover occurs to
bring back the MMM process on that control node.
Note: Only the active control node and the redundant control node’s monitor daemon
monitors each other, while the active control node also monitors non control node’s
monitoring daemon.
If the monitoring daemon has hung or been killed and there is no redundant control node,
restart IBM Spectrum Archive EE to start a new monitoring daemon.
As for the rest of the components, currently there are no automatic recovery actions that can
be performed. If GPFS, HSM, rpcbind, or rsyslog are having problems, the issues can be
viewed by using the eeadm node list command.
Example 6-36 shows output from running eeadm node list when one node has not started
rpcbind. To correct this error, start rpcbind on the designated node and EE will refresh itself
and the node will become available.
Spectrum Archive EE service (MMM) for library libb fails to start or is not
running on daito Node ID:1
Problem Detected:
Node ID Error Modules
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 153
1 MMM; rpcbind;
Node ID State Node IP Drives Ctrl Node Library Node Group Host
Name
2 available 9.11.244.62 2 yes(active) liba G0 nara
In addition to the automatic failover, there is also an option to perform a manual failover by
running the eeadm cluster failover command. This command is used to fail over the MMM
process to a redundant control node. This command is only available for use when a
redundant control node exists. Example 6-37 shows output from running the eeadm cluster
failover command.
After tape cartridges are added through the I/O station, or after they are inserted directly into
the tape library, you might have to run an eeadm library rescan command. First, run the
eeadm tape list command. If the tape cartridges are missing, run the eeadm library rescan
command, which synchronizes the data for these changes between the IBM Spectrum
Archive EE system and the tape library.
This process occurs automatically. However, if the tape does not appear within the eeadm
tape list command output, you can force a rebuild of the inventory (synchronization of IBM
Spectrum Archive EE inventory with the tape library’s inventory).
154 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 6-38 Run the eeadm tape list command to check whether a tape cartridge must be synchronized
[root@mikasa1 ~]# eeadm tape list -l lib_saitama
Tape ID Status State Usable(GiB) Used(GiB) Available(GiB) Reclaimable% Pool Library Location Task ID
JCA224JC ok appendable 6292 0 6292 0% pool1 lib_saitama homeslot -
JCC093JC ok appendable 6292 496 5796 0% pool1 lib_saitama homeslot -
JCB745JC ok appendable 6292 0 6292 0% pool2 lib_saitama homeslot -
4. Repeating the eeadm tape list command shows that the inventory was corrected, as
shown in Example 6-40.
5. If necessary, move the tape cartridge from the I/O station to a storage slot by running the
eeadm tape move command (see Example 6-41) with the -L homeslot option. The example
also requires a -l option because of multiple tape libraries.
6. Add the tape cartridge to a tape cartridge pool. If the tape cartridge contains actual data to
be added to LTFS, you must import it first before you add it. Run the eeadm tape import
command to add the tape cartridge into the IBM Spectrum Archive EE library and import
the files on that tape cartridge into the IBM Spectrum Scale namespace as stub files.
If you have no data on the tape cartridge (but it is already formatted for LTFS), add it to a
tape cartridge pool by running the eeadm tape assign command.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 155
Scratch cartridge
To add a scratch cartridge to the IBM Spectrum Archive EE system, complete the following
steps:
1. Insert the tape cartridge into the I/O station.
2. Synchronize the data in the IBM Spectrum Archive EE inventory with the tape library by
running the eeadm library rescan command, as shown in Example 6-39 on page 155.
3. If necessary, move the tape cartridge from the I/O station to a storage slot by running the
eeadm tape move command with the -L homeslot option, as shown in Example 6-41 on
page 155.
4. The eeadm tape assign command automatically formats tapes when assigning them to a
pool. Use the -f or --format option only when the user is aware that the tapes still contain
files and are no longer needed. For example, Example 6-42 shows the output of the eeadm
tape assign command.
Note: Media optimization will run on the each initial assignment starting from LTO 9, which
may take some more time. Refer to 7.29, “LTO 9 Media Optimization” on page 272 for more
information.
For more information about other formatting options, see 6.8.3, “Formatting tape cartridges”
on page 158.
Example 6-43 Error message when removing a tape cartridge from a pool with migrated/saved files
[root@mikasa1 ~]# eeadm tape unassign JCC093JC -p pool1 -l lib_saitama
2019-01-07 15:10:36 GLESL700I: Task tape_unassign was created successfully, task
id is 6970.
2019-01-07 15:10:36 GLESL357E: Tape JCC093JC has migrated files or saved files. It
has not been unassigned from the pool.
156 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
However, you can remove an empty tape cartridge from one tape cartridge pool and add it to
another tape cartridge pool, as shown in Example 6-44.
Example 6-44 Remove an empty tape cartridge from one tape cartridge pool and add it to another
[root@mikasa1 ~]# eeadm tape unassign JCA561JC -p pool1 -l lib_saitama
2019-01-07 15:11:26 GLESL700I: Task tape_unassign was created successfully, task id is
6972.
2019-01-07 15:11:26 GLESM399I: Removing tape JCA561JC from pool pool1 (Normal).
2019-01-07 15:11:26 GLESL359I: Unassigned tape JCA561JC from pool pool1 successfully.
Before you remove a tape cartridge from one tape cartridge pool and add it to another tape
cartridge pool, reclaim the tape cartridge to ensure that no files remain on the tape when it is
removed. For more information, see 6.17, “Reclamation” on page 200.
Example 6-45 Move a tape cartridge from a tape drive to its homeslot
[root@kyoto prod]# eeadm tape move D00369L5 -p pool1 -L homeslot
2019-01-07 15:47:26 GLESL700I: Task tape_move was created successfully, task id is
18843.
2019-01-07 15:49:14 GLESL103I: Tape D00369L5 is moved successfully.
Example 6-46 Move a tape cartridge to the ieslot after an offline operation
[root@mikasa1 ~]# eeadm tape offline JCA561JC -p pool2 -l lib_saitama
2019-01-07 15:50:17 GLESL700I: Task tape_offline was created successfully, task id
is 6976.
2019-01-07 15:50:17 GLESL073I: Offline export of tape JCA561JC has been requested.
2019-01-07 15:51:51 GLESL335I: Updated offline state of tape JCA561JC to offline.
The move can be between homeslot and ieslot or tape drive and homeslot. If the tape
cartridge belongs to a tape cartridge pool and online (not in the Offline state), the request to
move it to the ieslot fails. After a tape cartridge is moved to ieslot, the tape cartridge cannot
be accessed from IBM Spectrum Archive EE. If the tape cartridge contains migrated files, the
tape cartridge should not be moved to ieslot without first exporting or offlining the tape
cartridge.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 157
A tape cartridge in ieslot cannot be added to a tape cartridge pool. Such a tape cartridge
must be moved to home slot before adding it.
If the tape cartridge is already formatted for IBM Spectrum Archive EE and contains file
objects, the format fails, as shown in Example 6-47.
When the formatting is requested, IBM Spectrum Archive EE attempts to mount the target
medium to obtain the medium condition. The medium is formatted if the mount command finds
any of the following conditions:
The medium was not yet formatted for LTFS.
The medium was previously formatted for LTFS and has no data written.
The medium has an invalid label.
Labels in both partitions do not have the same value.
If none of these conditions are found, the format fails. If the format fails because files exist on
the tape, the user should add the tape to their designated pool by using the eeadm tape
import command. If the user no longer requires what is on the tape, the -f/--format option
can be added to the eeadm tape assign command to force a format.
Example 6-48 shows a tape cartridge being formatted by using the -f option.
158 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2019-01-08 09:30:33 GLESL360I: Assigned tape JCC130JC to pool pool1 successfully.
2019-01-08 09:30:52 GLESL087I: Tape JCC075JC successfully formatted.
2019-01-08 09:30:52 GLESL360I: Assigned tape JCC075JC to pool pool1 successfully.
When multiple format tasks are submitted, IBM Spectrum Archive EE uses all available drives
with the ‘g’ drive attribute for the format tasks, which are done in parallel.
The fastest method is to manage the pool and identify data in tape cartridges that has passed
the retention period. Customers can then remove the tape cartridge and add to a new pool by
reformatting the entire tape cartridge. This approach saves time, but customers need to be
certain that the tape cartridge does not have any active data.
To be sure that a tape cartridge is format-ready, this section uses the -E option, to the eeadm
tape unassign command. When ran, this command checks whether the tape cartridge
contains any active data. For example, if all files that were migrated on the tape has been
made resident, there will be no active data on tape. If there is no active data, or if all files in
the tape cartridge have already been deleted in GPFS, the command determines that the
tape cartridge is effectively empty and removes the tape cartridge from the pool. If the tape
cartridge still has active data, the command will not remove it. No reconciliation command
is necessary before this command.
Example 6-50 Removing tape cartridge from pool with active file check
[root@mikasa1 prod]# eeadm tape unassign JCB350JC -p test2 -l lib_saitama -E
2019-01-08 10:20:07 GLESL700I: Task tape_unassign was created successfully,
task id is 7002.
2019-01-08 10:20:09 GLESM399I: Removing tape JCB350JC from pool test2
(Empty_Force).
2019-01-08 10:20:09 GLESL572I: Unassign tape JCB350JC from pool test2
successfully. Format the tape when assigning it back to a pool.
5. If the tape cartridge has a valid, active file, the check routine aborts on the first hit and
goes on to the next specified tape cartridge. The command will not remove the tape
cartridge from the pool.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 159
Example 6-51 shows the output of the eeadm tape unassign -E command with a tape
which contains active files.
Example 6-51 Tape cartridges containing inactive data are removed from the pool
[root@mikasa1 prod]# eeadm tape unassign JCB350JC -p test2 -l lib_saitama -E
2019-01-08 10:17:15 GLESL700I: Task tape_unassign was created successfully,
task id is 6998.
2019-01-08 10:17:16 GLESL357E: Tape JCB350JC has migrated files or saved files.
It has not been unassigned from the pool.
The active file check applies to all data types that the current IBM Spectrum Archive EE might
store to a tape cartridge:
Normal migrated files
Saved objects such as empty directory and link files
Another approach is to run mmapplypolicy to list all files that have been migrated to the
designated tape cartridge ID. However, if the IBM Spectrum Scale file system has over
1 billion files, the mmapplypolicy scan might take a long time.
After mounting the library, the user can run eeadm commands to manage the library and to
correct a problem if one occurs.
To remove a tape drive from the library, complete the following steps:
1. Remove the tape drive from the IBM Spectrum Archive EE inventory by running the eeadm
drive unassign command, as shown in Example 6-52. A medium in the tape drive is
automatically moved to the home slot (if one exists).
160 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
[root@yurakucho ~]# eeadm drive list -l lib1
Drive S/N Status State Type Role Library Node ID Tape Node Group
Task ID
00078DF0CF ok not_mounted LTO8 mrg lib1 1 - G0 -
00078DF0D3 ok not_mounted LTO8 mrg lib1 1 - G0 -
00078DF0DC ok not_mounted LTO8 mrg lib1 1 - G0 -
00078DF0E2 info unassigned NONE --- lib1 - - - -
1013000508 ok not_mounted LTO8 mrg lib1 1 - G0 -
For more information about how to remove tape drives, see the IBM Documentation website
for your IBM tape library.
Optionally, drive attributes can be set when adding a tape drive. Drive attributes are the
logical OR of the attributes: migrate(4), recall(2), and generic(1). If the individual attribute
is set, any corresponding tasks on the task queue can be run on that drive. The drive
attributes can be specified using the -r option and must be a decimal number or the
combination of the letters m, r, and g.
In Example 6-53, 6 is the logical OR of migrate(4) and recall(2), so migration tasks and
recall tasks can be performed on this drive. For more information, see 6.20, “Drive Role
settings for task assignment control” on page 207.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 161
6.9 Tape storage pool management
This section describes how to use the eeadm pool command to manage tape cartridge pools
with IBM Spectrum Archive EE.
Permissions: Managing tape cartridge pools by running the eeadm pool command
requires root user permissions.
To perform file migrations, it is first necessary to create and define tape cartridge pools, which
are the targets for migration. It is then possible to add or remove tape cartridges to or from the
tape cartridge pools.
Consider the following rules and recommendations for tape cartridge pools:
Before adding tape cartridges to a tape cartridge pool, the tape cartridge must first be in
the homeslot of the tape library. For more information about moving to the homeslot, see
6.8.2, “Moving tape cartridges” on page 156.
Multiple tasks can be performed in parallel when more than one tape cartridge is defined
in a tape cartridge pool. Have multiple tape cartridges in each tape cartridge pool to
increase performance.
The maximum number of drives in a node group that is used for migration for a particular
tape cartridge pool can be limited by setting the mountlimit attribute for the tape cartridge
pool. The default is 0, which is unlimited. For more information about the mountlimit
attribute, see 7.2, “Maximizing migration performance with redundant copies” on
page 235.
After a file is migrated to a tape cartridge pool, it cannot be migrated again to another tape
cartridge pool before it is recalled.
When tape cartridges are removed from a tape cartridge pool but not exported from IBM
Spectrum Archive EE, they are no longer targets for migration or recalls.
When tape cartridges are exported from IBM Spectrum Archive EE system by running the
eeadm tape export command, they are removed from their tape cartridge pool and the
files are not accessible for recall.
You create tape cartridge pools by using the create option of the eeadm pool command. For
example, the command that is shown in Example 6-54 creates the tape cartridge pool named
MPEGpool.
For single tape library systems, the -l option (library name) can be omitted. For two tape
library systems, the -l option is used to specify the library name.
162 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
For single node group systems, the -g option (node group) can be omitted. For multiple node
group systems, the -g option is used to specify the node group.
The default tape cartridge pool type is a regular pool. If a WORM pool is wanted, supply the
--worm physical option.
The pool names are case-sensitive and can be duplicated in different tape libraries. No
informational messages are shown at the successful completion of the command. However,
you can confirm that the pool was created by running the eeadm pool list command.
For single tape library systems, the -l option (library name) can be omitted. For two tape
library systems, the -l option is used to specify the library name.
When deleting a tape cartridge pool, the -g option (node group) can be omitted.
No informational messages are shown after the successful completion of the command.
Important: If the tape cartridge pool contains tape cartridges, the tape cartridge pool
cannot be deleted until the tape cartridges are removed.
You cannot use IBM Spectrum Archive EE to delete a tape cartridge pool that still contains
data.
To allow the deletion of the tape cartridge pool, you must remove all tape cartridges from it by
running the eeadm tape unassign command, as described in 6.8.2, “Moving tape cartridges”
on page 156.
To enable the pool capacity monitoring feature, use the eeadm pool set command with the
lowspacewarningenable yes and the nospacewarningenable yes option, then set a
threshold limit for the pool with the lowspacewarningthreshold option. The value for the
lowspacewarningthreshold must be an integer and is in TiB. The pool capacity monitor
thread checks each set pool for a low capacity every 30 minutes, and sends a trap every 24
hours.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 163
Run the eeadm pool show <pool_name> command to view the current attributes for your pools.
Example 6-56 shows the output of the attributes of a pool from the eeadm pool show
command.
The lowspacewarningthreshold attribute value is set in TiB. To modify the attributes in each
pool, use the eeadm pool set <pool_name> -a <attribute> -v <value> command.
164 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
With lowspacewarningthreshold set to 30 TiB, when pool1’s capacity drops below 30 TiB, a
trap will be sent to the user when the next check cycle occurs. Example 6-58 shows the traps
generated when pool1’s capacity drops below 30 TiB.
Example 6-58 Traps sent when pool capacity is below the set threshold
6.11 Migration
The migration process is the most significant reason for using IBM Spectrum Archive EE.
Migration is the movement of files from IBM Spectrum Scale (on disk) to LTFS tape cartridges
in tape cartridge pools, which leaves behind a small stub file on the disk. This process has the
obvious effect of reducing the usage of file system space of IBM Spectrum Scale. You can
move less frequently accessed data to lower-cost, lower-tier tape storage from where it can
be easily recalled.
IBM Spectrum Scale policies are used to specify files in the IBM Spectrum Scale namespace
(through a GPFS scan) to be migrated to the LTFS tape tier. For each specified GPFS file, the
file content, GPFS path, and user-defined extended attributes are stored in LTFS so that they
can be re-created at import. Empty GPFS directories or files are not migrated.
In addition, it is possible to migrate an arbitrary list of files directly by running the eeadm
migrate command. This task is done by specifying the file name of a scan list file that lists the
files to be migrated and specifying the designated pools as command options.
Important: Running the IBM Spectrum Protect for Space Management dsmmigrate
command directly is not supported.
To migrate files, the following configuration and activation prerequisites must be met:
Ensure that the MMM service is running on an LTFS node. For more information, see
6.2.4, “IBM Spectrum Archive EE” on page 140.
Ensure that one or more storage pools are created and each has one or more assigned
tapes. For more information, see 6.9.1, “Creating tape cartridge pools” on page 162.
Ensure that space management is turned on. For more information, see 6.2.3,
“Hierarchical Space Management” on page 138.
Activate one of the following mechanisms to trigger migration:
– Automated IBM Spectrum Scale policy-driven migration that uses thresholds.
– Manual policy-based migration by running the mmapplypolicy command.
– Manual migration by running the eeadm migrate command and a prepared list of files
and tape cartridge pools.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 165
IBM Spectrum Archive EE uses a semi-sequential fill policy for tapes that enables multiple
files to be written in parallel by using multiple tape drives within the tape library. Tasks are put
on the queue and the scheduler looks at the queue to decide which tasks should be run. If
one tape drive is available, all of the migration goes on one tape cartridge. If there are three
tape drives available, the migrations are spread among the three tape drives. This
configuration improves throughput and is a more efficient usage of tape drives.
IBM Spectrum Archive EE internally groups files into file lists and schedules these lists on the
task queue. The lists are then distributed to available drives to perform the migrations.
The grouping is done by using two parameters: A total file size and a total number of files. The
default settings for the file lists are 20 GB or 20,000 files. This requirement means that a file
list can contain either 20 GB of files or 20,000 number of files, whichever fills up first, before
creating a new file list. For example, if you have 10 files to migrate and each file is 10 GB in
size, then when migration is kicked off, IBM Spectrum Archive EE internally generates five file
lists containing two files each because the two files reach the 20 GB limit that a file list can
have. It then schedules those file lists to the task queue for available drives.
For more information about performance references, see 3.7.4, “Performance” on page 65.
Note: Files that are recently created need to wait two minutes before being migrated.
Otherwise, the migrations will fail.
Example 6-59 shows the output of running the mmapplypolicy command that uses a policy file
called sample_policy.txt.
166 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
FALSE
OR PATH_NAME LIKE '/ibm/gpfs/.ltfsee/%'
OR PATH_NAME LIKE '%/.SpaceMan/%'
)
) OR ((
FALSE
)
))
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 167
6.11.1 Managing file migration pools
A file can be migrated to one pool or to multiple pools if replicas are configured. However,
after the file is in the migrated state, it cannot be migrated again to other tape cartridge pools
before it is recalled and made resident again by using the eeadm recall command with the
--resident option. For more information about creating replicas, see 6.11.4, “Replicas and
redundant copies” on page 177. Recalling the file into resident state invalidates the LTFS
copy from the reconcile and export perspective.
Automated IBM Spectrum Scale policy-driven migration is a standard IBM Spectrum Scale
migration procedure that allows file migration from IBM Spectrum Scale disk pools to external
pools. IBM Spectrum Archive EE is configured as an external pool to IBM Spectrum Scale by
using policy statements.
After you define an external tape cartridge pool, migrations or deletion rules can refer to that
pool as a source or target tape cartridge pool. When the mmapplypolicy command is run and
a rule dictates that data should be moved to an external pool, the user-provided program that
is identified with the EXEC clause in the policy rule starts. That program receives the following
arguments:
The command to be run. IBM Spectrum Scale supports the following subcommands:
– LIST: Provides arbitrary lists of files with no semantics on the operation.
– MIGRATE: Migrates files to external storage and reclaims the online space that is
allocated to the file.
– PREMIGRATE: Migrates files to external storage, but does not reclaim the online
space.
– PURGE: Deletes files from both the online file system and the external storage.
– RECALL: Recall files from external storage to the online storage.
– TEST: Tests for presence and operation readiness. Returns zero for success and
returns nonzero if the script should not be used on a specific node.
The name of a file that contains a list of files to be migrated.
Important: IBM Spectrum Archive EE supports only the LIST, MIGRATE, PREMIGRATE,
and RECALL subcommands.
Any optional parameters that are specified with the OPTS clause in the rule. These optional
parameters are not interpreted by the IBM Spectrum Scale policy engine, but by the
method that IBM Spectrum Archive EE uses to pass the tape cartridge pools to which the
files are migrated.
To set up automated IBM Spectrum Scale policy-driven migration to IBM Spectrum Archive
EE, you must configure IBM Spectrum Scale to be managed by IBM Spectrum Archive EE. In
addition, a migration callback must be configured.
168 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Callbacks are provided primarily as a method for system administrators to take notice when
important IBM Spectrum Scale events occur. It registers a user-defined command that IBM
Spectrum Scale runs when certain events occur. For example, an administrator can use the
low disk event callback to inform system administrators when a file system is getting full.
The migration callback is used to register the policy engine to be run if a high threshold in a
file system pool is met. For example, after your pool usage reaches 80%, you can start the
migration process. You must enable the migration callback by running the mmaddcallback
command.
In the mmaddcallback command in Example 6-60, the --command option points to a sample
script file called /usr/lpp/mmfs/bin/mmapplypolicy. Before you run this command, you must
ensure that the specified sample script file exists. The --event option registers the events for
which the callback is configured, such as the “low disk space” events that are in the command
example.
For more information about how to create and set a fail-safe policy, see 7.10, “Use cases for
mmapplypolicy” on page 244.
After the file system is configured to be managed by IBM Spectrum Archive EE and the
migration callback is configured, a policy can be set up for the file system. The placement
policy that defines the initial placement of newly created files and the rules for placement of
restored data must be installed into IBM Spectrum Scale by using the mmchpolicy command.
If an IBM Spectrum Scale file system does not have a placement policy installed, all the data
is stored in the system storage pool.
You can define the file management rules and install them in the file system together with the
placement rules by running the mmchpolicy command. You also can define these rules in a
separate file and explicitly provide them to the mmapplypolicy command by using the -P
option. The latter option is described in 6.11.3, “Manual migration” on page 173.
In either case, policy rules for placement or migration can be intermixed. Over the life of the
file, data can be migrated to a different tape cartridge pool any number of times, and files can
be deleted or restored.
Tip: Only one IBM Spectrum Scale policy, which can include one or more rules, can be set
up for a particular GPFS file system.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 169
After a policy is entered into a text file (such as policy.txt), you can apply the policy to the
file system by running the mmchpolicy command. You can check the syntax of the policy
before you apply it by running the command with the -I test option, as shown in
Example 6-61.
After you test your policy, run the mmchpolicy command without the -I test to set the policy.
After a policy is set for the file system, you can check the policy by displaying it with the
mmlspolicy command, as shown in Example 6-62. This policy migrates all files in groups of
20 GiB after the IBM Spectrum Archive disk space reaches a threshold of or above 80% in the
/ibm/glues/archive directory to tape.
define(
user_exclude_list,
PATH_NAME LIKE '/ibm/glues/0%'
OR NAME LIKE '%&%')
define(
user_include_list,
FALSE)
define(
exclude_list,
NAME LIKE 'dsmerror.log')
170 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
TO POOL 'Archive_files'
WHERE PATH_NAME LIKE '/ibm/glues/archive/%'
AND NOT (exclude_list)
AND (NOT (user_exclude_list) OR (user_include_list))
AND (is_migrated OR is_premigrated)
To ensure that a specified file system is migrated only once, run the mmapplypolicy command
with the --single-instance option. If this is not done, IBM Spectrum Archive EE attempts to
start another migration process every two minutes. This occurs in cases where migrations are
continuously called while the past migrations has not finished, because this state may appear
for the values to be still over the threshold.
As a preferred practice, the user should not use overlapping IBM Spectrum Scale policy rules
within different IBM Spectrum Scale policy files that select the same files for migration to
different tape cartridge pools. If a file is already migrated, later migration attempts fail, which
is the standard HSM behavior. However, this is not normally done, due to incorporating
thresholds.
Important: If a single IBM Spectrum Scale file system is used and the metadata directory
is stored in the same file system that is space-managed with IBM Spectrum Archive EE,
migration of the metadata directory must be prevented. The name of metadata directory is
<GPFS mount point>/.ltfsee/.
By combining the attributes of THRESHOLD and WEIGHT in IBM Spectrum Scale policies, you can
have a great deal of control over the migration process. When an IBM Spectrum Scale policy
is applied, each candidate file is assigned a weight (based on the WEIGHT attribute). All
candidate files are sorted by weight and the highest weight files are chosen to MIGRATE until
the low occupancy percentage (based on the THRESHOLD attribute) is achieved, or there are no
more candidate files.
Example 6-63 shows a policy that starts migration of all files when the file system pool named
“system” reaches 80% full (see the THRESHOLD attribute), and continues migration until the
pool is reduced to 60% full or less by using a weight that is based on the date and time that
the file was last accessed (refer to the ACCESS_TIME attribute). The file system usage is
checked every two minutes.
All files to be migrated must have more than 5 MB of disk space that is allocated for the file
(see the KB_ALLOCATED attribute). The migration is performed to an external pool, presented by
IBM Spectrum Archive EE (/opt/ibm/ltfsee/bin/eeadm), and the data that is migrated is sent
to the IBM Spectrum Archive EE tape cartridge pool named Tapepool1. In addition, this
example policy excludes some system files and directories.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 171
RULE EXTERNAL POOL 'ltfsee'
EXEC '/opt/ibm/ltfsee/bin/eeadm'
OPTS -p 'Tapepool1@liba' /* This is our pool in LTFS Enterprise Edition */
SIZE(20971520)
THRESHOLD(80,60)
WEIGHT(CURRENT_TIMESTAMP - ACCESS_TIME)
TO POOL 'ltfsee'
WHERE (KB_ALLOCATED > 5120)
AND NOT (exclude_list)
/* The following statement is the default placement rule that is required for a
system migration */
RULE 'default' set pool 'system'
In addition to monitoring the file system’s overall usage in Example 6-63, you can monitor
how frequently a file is accessed with IBM Spectrum Scale policies. A file’s access
temperature is an attribute for a policy that provides a means of optimizing tiered storage. File
temperatures are a relative attribute, which indicates whether a file is “hotter” or “colder” than
the others in its pool.
The policy can be used to migrate hotter files to higher tiers and colder files to lower. The
access temperature is an exponential moving average of the accesses to the file. As files are
accessed, the temperature increases. Likewise, when the access stops, the file cools. File
temperature is intended to optimize nonvolatile storage, not memory usage. Therefore, cache
hits are not counted. In a similar manner, only user accesses are counted.
The access counts to a file are tracked as an exponential moving average. A file that is not
accessed loses a percentage of its accesses each period. The loss percentage and period
are set through the configuration variables fileHeatLossPercent and
fileHeatPeriodMinutes. By default, the file access temperature is not tracked.
To use access temperature in policy, the tracking must first be enabled. To do this, set the
following configuration variables:
fileHeatLossPercent
The percentage (0 - 100) of file access temperature that is dissipated over the
fileHeatPeriodMinutes time. The default value is 10.
fileHeatPeriodMinutes
The number of minutes that is defined for the recalculation of file access temperature. To
turn on tracking, fileHeatPeriodMinutes must be set to a nonzero value from the default
value of 0. You use WEIGHT(FILE_HEAT) with a policy MIGRATE rule to prioritize migration by
file temperature.
172 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Note: If the updating of the file access time (atime) is suppressed or if relative atime
semantics are in effect, proper calculation of the file access temperature might be
adversely affected.
These examples provide only an introduction to the wide range of file attributes that migration
can use in IBM Spectrum Scale policies. IBM Spectrum Scale provides a range of other policy
rule statements and attributes to customize your IBM Spectrum Scale environment, but a full
description of all these is outside the scope for this publication.
For syntax definitions for IBM Spectrum Scale policy rules, which correspond to constructs in
this script (such as EXEC, EXTERNAL POOL, FROM POOL, MIGRATE, RULE, OPTS, THRESHOLD, TO POOL,
WEIGHT, and WHERE), see the information about policy rule syntax definitions in IBM Spectrum
Scale: Administration Guide at IBM Documentation.
Also, see 7.4, “Setting mmapplypolicy options for increased performance” on page 237.
For more information about IBM Spectrum Scale SQL expressions for policy rules, which
correspond to constructs in this script (such as CURRENT_TIMESTAMP, FILE_SIZE,
MISC_ATTRIBUTES, NAME, and PATH_NAME), see the information about SQL expressions for
policy rules in IBM Spectrum Scale: Administration Guide at IBM Documentation.
Using mmapplypolicy
This section describes how to manually start file migration while using an IBM Spectrum
Scale policy file for file selection.
You can apply a manually created policy by manually running the mmapplypolicy command,
or by scheduling the policy with the system scheduler. You can have multiple different policies,
which can each include one or more rules. However, only one policy can be run at a time.
You can accomplish manual file migration for an IBM Spectrum Scale file system that is
managed by IBM Spectrum Archive EE by running the mmapplypolicy command. This
command runs a policy that selects files according to certain criteria, and then passes these
files to IBM Spectrum Archive EE for migration. As with automated IBM Spectrum Scale
policy-driven migrations, the name of the target IBM Spectrum Archive EE tape cartridge pool
is provided as the first option of the pool definition rule in the IBM Spectrum Scale policy file.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 173
The following phases occur when the mmapplypolicy command is started:
1. Phase 1: Selecting candidate files
In this phase of the mmapplypolicy job, all files within the specified GPFS file system
device (or below the input path name) are scanned. The attributes of each file are read
from the file’s GPFS inode structure.
2. Phase two: Choosing and scheduling files
In this phase of the mmapplypolicy job, some or all of the candidate files are chosen.
Chosen files are scheduled for migration, accounting for the weights and thresholds that
are determined in phase one.
3. Phase three: Migrating and premigrating files
In the third phase of the mmapplypolicy job, the candidate files that were chosen and
scheduled by the second phase are migrated or premigrated, each according to its
applicable rule.
For more information about the mmapplypolicy command and other information about IBM
Spectrum Scale policy rules, see the IBM Spectrum Scale: Administration Guide, which is
available at IBM Documentation.
174 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Testing an IBM Spectrum Scale policy
Example 6-64 shows a mmapplypolicy command that tests, but does not apply, an IBM
Spectrum Scale policy by using the testpolicy policy file.
THRESHOLD(50,0)
TO POOL 'COPY_POOL'
WHERE FILE_SIZE > 5242880
AND NAME LIKE ‘%.IMG’
AND ((NOT MISC_ATTRIBUTES LIKE '%M%') OR (MISC_ATTRIBUTES LIKE '%M%' AND
MISC_ATTRIBUTES NOT LIKE '%V%'))
AND NOT ((PATH_NAME LIKE '/ibm/gpfs/.ltfsee/%' OR PATH_NAME LIKE '%/.SpaceMan/%'))
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 175
The policy in Example 6-64 on page 175 is configured to select files that have the file
extension .IMG for migration to the IBM Spectrum Archive EE tape cartridge pool named
copy_ltfsml1 in library name lib_ltfsml1 if the usage of the /ibm/gpfs file system exceeds
50% for any .IMG file that exceeds 5 MB.
Using eeadm
The eeadm migrate command requires a migration list file that contains a list of files to be
migrated with the name of the target tape cartridge pool. Unlike migrating files by using IBM
Spectrum Scale policy, it is not possible to use wildcards in place of file names. The name
and path of each file to be migrated must be specified in full. The file must be in the following
format:
/ibm/glues/file1.mpeg
/ibm/glues/file2.mpeg
176 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The syntax for a cron job is m h dom mon dow command. In this syntax, m stands for minutes,
h stands for hours, dom stands for day of month, mon stands for month, and dow stands for day
of week. The hour parameter is in a 24-hour period, so 0 represents midnight and 12
represents noon.
Example 6-66 shows how to start the crond process and create a single cron job that
performs migrations every six hours.
Example 6-66 Creating a cron job for migrations to run every 6 hours
[root@ltfseesrv ~]# /etc/rc.d/init.d/crond start
Starting crond: [ OK ]
Overview
IBM Spectrum Archive EE enables the creation of a replica of each IBM Spectrum Scale file
during the migration process. The purpose of the replica function is to enable creating
multiple LTFS copies of each GPFS file during migration that can be used for disaster
recovery, including across two tape libraries at two different locations.
The first replica is the primary copy, and more replicas are called redundant copies.
Redundant copies must be created in tape cartridge pools that are different from the pool of
the primary copy and from the pools of other redundant copies. Up to two redundant copies
can be created (for a total of three copies of the file on various tapes).
The tape cartridge where the primary copy is stored and the tape cartridges that contain the
redundant copies are referenced in the GPFS inode with an IBM Spectrum Archive EE
DMAPI attribute. The primary copy is always listed first.
For transparent recalls such as double-clicks of a file or through application reads, IBM
Spectrum Archive EE always performs the recall by using the primary copy tape. The primary
copy is the first tape cartridge pool that is defined by the migration process. If the primary
copy tape cannot be accessed, including recall failures, then IBM Spectrum Archive EE
automatically tries the recall task again by using the remaining replicas if they are available
during the initial migration process. This automatic retry operation is transparent to the
transparent recall requester.
For selective recalls initiated by the eeadm recall command, an available copy is selected
from the available replicas and the recall task is generated against the selected tape
cartridge. There are no retries. The selection is based on the available copies in the tape
library, which is supplied by the -l option in a two-tape library environment.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 177
When a migrated file is recalled for a write operation or truncated, the file is marked as
resident and the pointers to tape are dereferenced. The remaining copies are no longer
referenced and are removed during the reconciliation process. In the case where a migrated
file is truncated to 0, it does not generate a recall from tape. The truncated 0 file is marked as
resident only.
Redundant copies are written to their corresponding tape cartridges in the IBM Spectrum
Archive EE format. These tape cartridges can be reconciled, exported, reclaimed, or
imported by using the same commands and procedures that are used for standard migration
without replica creation.
If an IBM Spectrum Scale scan is used and you use a scan policy file to specify files for
migration, you must modify the OPTS line of the policy file to specify the tape cartridge pool for
the primary replica and different tape cartridge pools for each redundant copy. The tape
cartridge pool for the primary replica (including primary library) is listed first, followed by the
tape cartridge pools for each copy (including a secondary library), as shown in Example 6-67.
A pool cannot be listed more than once in the OPTS line. If a pool is listed more than once per
line, the file is not migrated. Example 6-67 shows the OPTS line in a policy file, which makes
replicas of files in two tape cartridge pools in a single tape library.
Example 6-67 Extract from IBM Spectrum Scale policy file for replicas
OPTS '-p PrimPool@PrimLib CopyPool@PrimLib'
For more information about IBM Spectrum Scale policy files, see 6.11.2, “Threshold-based
migration” on page 168.
If you are running the eeadm migrate (or eeadm premigrate) command, a scan list file must be
passed along with the designated pools to which the user wants to migrate the files. This
process can be done by calling eeadm migrate <inputfile> -p <list_of_pools> [OPTIONS]
(or eeadm premigrate).
Example 6-68 shows what the scan list looks like when selecting files to migrate.
Example 6-69 shows how one would run a manual migration using the eeadm migrate
command on the scan list from Example 6-68 to two tapes.
178 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2019-01-08 15:52:39 GLESM897I: Starting the stage 2 of 3 for migration task 18866
(copying the files to 2 pools).
2019-01-08 15:53:26 GLESM898I: Starting the stage 3 of 3 for migration task 18866
(changing the state of files on disk).
2019-01-08 15:53:26 GLESL038I: Migration result: 2 succeeded, 0 failed, 0
duplicate, 0 duplicate wrong pool, 0 not found, 0 too small to qualify for
migration, 0 too early for migration.
IBM Spectrum Archive EE attempts to create redundant copies as efficiently as possible with
a minimum number of mount and unmount steps. For example, if all tape drives are loaded
with tape cartridges that belong only to the primary copy tape cartridge pool, data is written to
them before IBM Spectrum Archive EE begins loading the tape cartridges that belong to the
redundant copy tape cartridge pools. For more information, see 3.7, “Sizing and settings” on
page 61.
By monitoring the eeadm task list and the eeadm task show command as the migration is
running, you can observe the status of the migration task, as shown in Example 6-70.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 179
Completed Time: Tue Jan 8 16:12:14 2019 (-0700)
Workload: 2 files. 2 replicas.
4566941 bytes to copy. 1 copy tasklets on pool2@lib_saitama.
4566941 bytes to copy. 1 copy tasklets on test2@lib_saitama.
Progress: -
1/1 copy tasklets completed on pool2@lib_saitama.
1/1 copy tasklets completed on test2@lib_saitama.
Result Summary: 2 succeeded, 0 failed, 0 duplicate, 0 duplicate wrong pool,
0 not found, 0 too small, 0 too early.
(GLESM899I) All files have been successfully copied on
pool2/lib_saitama.
(GLESM899I) All files have been successfully copied on
test2/lib_saitama.
For more information and command syntax, see the eeadm migrate command in 6.11,
“Migration” on page 165.
Considerations
Consider the following points when replicas are used:
Redundant copies must be created in different tape cartridge pools. The pool of the
primary replica must be different from the pool for the first redundant copy, which, in turn,
must be different from the pool for the second redundant copy.
The migration of a premigrated file does not create replicas.
If offsite tapes are required, redundant copies can be exported out of the tape library and
shipped to an offsite location after running the eeadm tape export or eeadm tape offline,
depending on how the data should be kept. A second option would be to create the redundant
copy in a different tape library.
Data on older generation media can be moved as a whole within a pool or specific tapes can
be chosen in chunks. Example 6-71 shows how users can use the eeadm tape datamigrate
command to migrate their data from an older generation pool to a newer one.
180 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2021-03-24 04:23:29 GLESL677I: Files on tape JCA561JC will be copied to tape
JCB370JC.
2021-03-24 04:28:55 GLESL081I: Tape JCA561JC successfully reclaimed as datamigrate
process, formatted, and unassigned from tape pool test3.
2021-03-24 04:28:55 GLESL080I: Reclamation complete as the datamigrate command. 1
tapes reclaimed, 1 tapes unassigned from the tape pool.
For the eeadm tape datamigrate command syntax, see “The eeadm <resource type>
<action> --help command” on page 317.
Note: If the tapes are not specified for eeadm tape reclaim, the tapes being reclaimed
remain in the source pool. If the tapes are specified with the eeadm tape datamigrate
command, those tapes will be removed from the source pool after the reclamation
completes.
In addition to the pool-to-pool data migration, users can perform in-pool data migration by
configuring the pool settings and modifying the media_restriction or format type to make
older generation media become append_fenced. By doing so, in addition to securing future
migration tasks go to newer generation media this also enables users to run reclamation on
those older generation tape cartridges and have assurance that the data is reclaimed onto
the new generation medias.
When changing the media_restriction attribute of the pool, the format type is also
automatically updated to the highest generation drive available to the cluster. The format
attribute is automatically updated only after each time the media_restriction is modified. If
new drive and media generations are added to a cluster with media_restriction already set,
users are expected to update format or media_restriction manually to support new media
generations.
Example 6-72 on page 181 shows the automatic format update when changing the
media_restriction attribute from JC to JE when there are TS1160, TS1155, and TS1150
drives.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 181
[root@mikasa1 prod]# eeadm pool show test3 -l lib_saitama
Attribute Value
poolname test3
poolid 99096daa-0beb-4791-b24f-672804a56440
devtype 3592
mediarestriction JE
format 60F (0x57)
worm no (0)
nodegroup G0
fillpolicy Default
owner System
mountlimit 0
lowspacewarningenable yes
lowspacewarningthreshold 0
nospacewarningenable yes
mode normal
All tapes within test3 that fall under the restriction of JE media and 60F format type are all
appendable tapes, and everything else is now append_fenced.
It is preferable that you do not use overlapping IBM Spectrum Scale policy rules within
different IBM Spectrum Scale policy files that can select the same files for migration to
different tape cartridge pools. If a file is already migrated, a later migration fails.
In this example, an attempt is made to migrate four files to tape cartridge pool pool2. Before
the migration attempt, tape JCB610JC, which is defined in a different tape cartridge pool
(pool1), already contains three of the four files. The state of the files on these tape cartridges
before the migration attempt is displayed by the eeadm file state command as shown in
Example 6-73.
Name: /ibm/gpfs/prod/fileB.ppt
State: migrated
ID: 11151648183451819981-3451383879228984073-1844785692-3068794-0
Replicas: 1
Tape 1: JCB610JC@pool1@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/fileC.ppt
State: migrated
182 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
ID: 11151648183451819981-3451383879228984073-373707969-3068783-0
Replicas: 1
Tape 1: JCB610JC@pool1@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/fileD.ppt
State: resident
The attempt to migrate the files to a different tape cartridge pool produces the results that are
shown in Example 6-74.
If the IBM Spectrum Archive EE log is viewed, the error messages that are shown in
Example 6-75 explain the reason for the failures.
Example 6-75 Migration errors reported in the IBM Spectrum Archive EE log file
2019-01-09T15:07:18.713348-07:00 saitama2 mmm[22592]: GLESM148E(00710): File
/ibm/gpfs/prod/fileA.ppt is already migrated and will be skipped.
2019-01-09T15:07:18.714413-07:00 saitama2 mmm[22592]: GLESM148E(00710): File
/ibm/gpfs/prod/fileB.ppt is already migrated and will be skipped.
2019-01-09T15:07:18.715196-07:00 saitama2 mmm[22592]: GLESM148E(00710): File
/ibm/gpfs/prod/fileC.ppt is already migrated and will be skipped.
The files on tape JCB610JC (fileA.ppt, fileB.ppt, and fileC.ppt) are already in storage pool
pool1. Therefore, the attempt to migrate them to storage pool pool2 produces a migration
result of Failed. Only the attempt to migrate the resident file fileD.ppt succeeds.
If the aim of this migration was to make redundant replicas of the four PPT files in the pool2
tape cartridge pool, the method that is described in 6.11.4, “Replicas and redundant copies”
on page 177 must be followed instead.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 183
Automated IBM Spectrum Scale policy-driven migration
To ensure that a specified GPFS file system is migrated only once, run the mmapplypolicy
command with the --single-instance option. The --single-instance option ensures that
multiple mmapplypolicy commands are not running in parallel because it can take longer than
two minutes to migrate a list of files to tape cartridges.
Tape format
For more information about the format of tapes that are created by the migration process, see
10.2, “Formats for IBM Spectrum Scale to IBM Spectrum Archive EE migration” on page 325.
Migration Policy
A migration policy is used to make your lives easier. When run, IBM Spectrum Scale performs
a scan of all candidate files in the IBM Spectrum Archive name space to be migrated onto
tape. This process saves the user lots of time because they do not need to manually search
their file system and find candidate files for migrations. This feature is especially important
when there are millions of files created. For use cases on migration policy, see 7.10, “Use
cases for mmapplypolicy” on page 244.
6.12 Premigration
A premigrated file is a file that the content is on both disk and tape. To change a file to a
premigrated state, you have two options:
Recalling migrated files:
a. The file initially is only on a disk (the file state is resident).
b. The file is migrated to tape by running eeadm migrate. After this migration, the file is a
stub on the disk (the file state is migrated) and the IDs of the tapes containing the
redundant copies are written to an IBM Spectrum Archive EE DMAPI attribute.
c. The file is recalled from tape by using a recall for read when a client attempts to read
from the file. The content of the file is on both disk and tape (the file state is
premigrated).
Premigrating files:
a. The file initially is only on disk (the file state is resident).
b. The file is premigrated to tape by running eeadm premigrate. The IDs of the tapes that
contain the redundant copies are written to an IBM Spectrum Archive EE DMAPI
attribute.
184 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
For the eeadm migrate command, each migrate task is achieved internally by splitting the
work into three steps:
1. Writing the content of the file to tapes, including redundant copies.
2. Writing the IDs of the tapes that contain the redundant copies of the file, which are written
to an IBM Spectrum Archive EE DMAPI attribute.
3. Stubbing the file on disk.
For premigration, step 3 is not performed. The omission of this step is the only difference
between premigration and migration.
The <input_file> file includes the list of non-empty regular files to be premigrated. Each line
of this file must end with either of the following format.
Each line ends with -- <filename> with a space before and after the double dash (the file
list record format of the mmapplypolicy command).
Each line contains a file name with an absolute path or a relative path that is based on the
working directory. This format is unavailable when you run the command with the
--mmbackup option.
Specified files are saved to the specified target tape cartridge pool. Optionally, the target tape
cartridge pool can be followed by up to two more tape cartridge pools (for redundant copies)
separated by commas.
The THRESHOLD clause can have the following parameters to control migration and
premigration:
THRESHOLD (high percentage, low percentage, premigrate percentage)
If no premigrate threshold is set with the THRESHOLD clause or a value is set greater than or
equal to the low threshold, then the mmapplypolicy command does not premigrate files. If the
premigrate threshold is set to zero, the mmapplypolicy command premigrates all files.
For example, the following rule premigrates all files if the storage pool occupancy is 0 - 30%.
When the storage pool occupancy is 30% or higher, files are migrated until the storage pool
occupancy drops below 30%. Then, it continues by premigrating all files:
RULE 'premig1' MIGRATE FROM POOL 'system' THRESHOLD (0,30,0) TO POOL 'ltfs'
The rule in the following example takes effect when the storage pool occupancy is higher than
50%. Then, it migrates files until the storage pool occupancy is lower than 30%, after which it
premigrates the remaining files:
RULE 'premig2' MIGRATE FROM POOL 'system' THRESHOLD (50,30,0) TO POOL 'ltfs'
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 185
The rule in the following example is configured so that if the storage pool occupancy is below
30%, it selects all files that are larger than 5 MB for premigration. Otherwise, when the
storage pool occupancy is 30% or higher, the policy migrates files that are larger than 5 MB
until the storage pool occupancy drops below 30%.
Then, it continues by premigrating all files that are larger than 5 MB:
RULE 'premig3' MIGRATE FROM POOL 'system' THRESHOLD (0,30,0) TO POOL 'ltfs' WHERE(
AND (KB_ALLOCATED > 5120))
The rule in the following example is the preferred rule when performing premigrations only. It
requires a callback to perform the stubbing. If the storage pools occupancy is below 100%, it
selects all files larger than 5 MB for premigration. If you set the threshold to 100% the storage
pools occupancy will never exceed this value, and so migrations will not be performed. In this
case, a callback is needed to run the stubbing. For an example of a callback, see 7.10.2,
“Creating active archive system policies” on page 245.
RULE ‘premig4’ MIGRATE FROM POOL ‘system’ THRESHOLD (0,100,0) TO POOL ‘ltfs’ WHERE
(FILE_SIZE > 5242880)
Because these file system objects do not have data, they cannot be processed by migration
or premigration. A new driver (called the save driver) was introduced to save these file system
objects to tape.
The following items (data and metadata that is associated with an object) are written and read
to and from tapes:
File data for non-empty regular files
Path and file name for all objects
Target symbolic name only for symbolic links
User-defined extended attributes for all objects except symbolic links
The following items are not written and read to and from tapes:
Timestamps
User ID and group ID
ACLs
To save these file system objects on tape, you have two options:
Calling the eeadm save command directly with a scan list file
An IBM Spectrum Scale policy with the mmapplypolicy command
6.13.1 Saving file system objects with the eeadm save command
The eeadm save command is used to save symbolic links, empty regular files, and empty
directories to tape. The command syntax is the same as the eeadm migrate or eeadm
premigrate commands. The following is the syntax of the eeadm save command:
eeadm save <inputfile> -p <list_of_pools> [OPTIONS]
186 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The <inputfile> file includes the list of file system objects (symbolic links, empty regular
files, and empty directories) to be saved. Each line of this file must follow either of the
following format.
Each line ends with -- <filename> with a space before and after the double dash (the file
list record format of the mmapplypolicy command).
Each line contains a file name with an absolute path or a relative path that is based on the
working directory. This format is unavailable when you run the command with the
--mmbackup option.
All file system objects are saved to the specified target tape cartridge pool. Optionally, the
target tape cartridge pool can be followed by up to two more tape cartridge pools (for
redundant copies) separated by commas.
A working policy sample of IBM Spectrum Scale list rules to save these file system objects
without data to tape can be found in the /opt/ibm/ltfsee/share/sample_save.policy file.
The only change that is required to the following sample policy file is the specification of the
cartridge pool (in blue colored letters).
These three list rules can be integrated into IBM Spectrum Scale policies. Example 6-76
shows the sample policy.
Example 6-76 Sample policy to save file system objects without data to tape
/*
Sample policy rules
to save
symbolic links,
empty directories and
empty regular files
*/
RULE
EXTERNAL LIST 'emptyobjects'
EXEC '/opt/ibm/ltfsee/bin/ltfseesave'
OPTS '-p sample_pool@sample_library'
define(DISP_XATTR,
CASE
WHEN XATTR($1) IS NULL
THEN '_NULL_'
ELSE XATTR($1)
END
)
RULE 'symoliclinks'
LIST 'emptyobjects'
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 187
DIRECTORIES_PLUS
/*
SHOW ('mode=' || SUBSTR(MODE,1,1) ||
' stime=' || DISP_XATTR('dmapi.IBMSTIME') ||
' ctime=' || VARCHAR(CHANGE_TIME) ||
' spath=' || DISP_XATTR('dmapi.IBMSPATH'))
*/
WHERE
( /* if the object is a symbolic link */
MISC_ATTRIBUTES LIKE '%L%'
)
AND
(
PATH_NAME NOT LIKE '%/.SpaceMan/%'
)
AND
(
( /* if the object has not been saved yet */
XATTR('dmapi.IBMSTIME') IS NULL
AND
XATTR('dmapi.IBMSPATH') IS NULL
)
OR
( /* if the object is modified or renamed after it was saved */
TIMESTAMP(XATTR('dmapi.IBMSTIME')) < TIMESTAMP(CHANGE_TIME)
OR
XATTR('dmapi.IBMSPATH') != PATH_NAME
)
)
RULE 'directories'
LIST 'emptyobjects'
DIRECTORIES_PLUS
/*
SHOW ('mode=' || SUBSTR(MODE,1,1) ||
' stime=' || DISP_XATTR('dmapi.IBMSTIME') ||
' ctime=' || VARCHAR(CHANGE_TIME) ||
' spath=' || DISP_XATTR('dmapi.IBMSPATH'))
*/
WHERE
( /* if the object is a directory */
MISC_ATTRIBUTES LIKE '%D%'
)
AND
(
PATH_NAME NOT LIKE '%/.SpaceMan'
AND
PATH_NAME NOT LIKE '%/.SpaceMan/%'
)
AND
(
( /* directory's emptiness is checked in the later processing */
/* if the object has not been saved yet */
XATTR('dmapi.IBMSTIME') IS NULL
AND
188 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
XATTR('dmapi.IBMSPATH') IS NULL
)
OR
( /* if the object is modified or renamed after it was saved */
TIMESTAMP(XATTR('dmapi.IBMSTIME')) < TIMESTAMP(CHANGE_TIME)
OR
XATTR('dmapi.IBMSPATH') != PATH_NAME
)
)
RULE 'emptyregularfiles'
LIST 'emptyobjects'
/*
SHOW ('mode=' || SUBSTR(MODE,1,1) ||
' stime=' || DISP_XATTR('dmapi.IBMSTIME') ||
' ctime=' || VARCHAR(CHANGE_TIME) ||
' spath=' || DISP_XATTR('dmapi.IBMSPATH'))
*/
WHERE
( /* if the object is a regular file */
MISC_ATTRIBUTES LIKE '%F%'
)
AND
(
PATH_NAME NOT LIKE '%/.SpaceMan/%'
)
AND
(
( /* if the size = 0 and the object has not been saved yet */
FILE_SIZE = 0
AND
XATTR('dmapi.IBMSTIME') IS NULL
AND
XATTR('dmapi.IBMSPATH') IS NULL
)
OR
( /* if the object is modified or renamed after it was saved */
FILE_SIZE = 0
AND
(
TIMESTAMP(XATTR('dmapi.IBMSTIME')) < TIMESTAMP(CHANGE_TIME)
OR
XATTR('dmapi.IBMSPATH') != PATH_NAME
)
)
)
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 189
6.14 Recall
In space management solutions, there are two different types of recall possibilities:
Transparent and selective recall processing. Both are possible with the current IBM Spectrum
Archive EE implementation.
Transparent recalls are initiated by an application that tries to read, write, or truncate a
migrated file while not being aware that it was migrated. The specific I/O request that initiated
the recall of the file is fulfilled, with a possible delay because the file data is not available
immediately (it is on tape).
Furthermore, with the default IBM Spectrum Archive EE settings, there is a limitation of up to
only 60 transparent recalls possible on the IBM Spectrum Archive EE task queue. A sixtyish
request appears only if one of the previous 60 transparent recall requests completes.
Therefore, the ordering can happen only on this small 60 transparent recall subset. It is up to
the software application to send the transparent recalls in parallel to have multiple transparent
recalls to run at the same time.
Selective recalls are initiated by users that are aware that the file data is on tape and they
want to transfer it back to disk before an application accesses the data. This action avoids
delays within the application that is accessing the corresponding files.
Contrary to transparent recalls, the performance objective for selective recalls is to provide
the best possible throughput for the complete set of files that is being recalled, disregarding
the response time for any individual file.
However, to provide reasonable response times for transparent recalls in scenarios where
recall of many files is in progress, the processing of transparent recalls are modified to have
higher priority than selective recalls. Selective recalls are performed differently than
transparent recalls, and so they do not have a limitation like transparent recalls.
Recalls have higher priority than other IBM Spectrum Archive EE operations. For example, if
there is a recall request for a file on a tape cartridge being reclaimed or for a file on the tape
cartridge being used as reclamation target, the reclamation task is stopped, the recall or
recalls from the tape cartridge that is needed for recall are served, and then the reclamation
resumes automatically.
Recalls also have higher priority over tape premigration processes. They are optimized
across tapes and optimized within a tape used for premigration activities. The recalls that are
in close proximity are given priority.
190 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
If you modify or truncate a recalled file, it becomes a resident file. The next time your file
system is reconciled, MMM marks the stored copy for deletion.
The order of selection from the replicas is always the same. The primary copy is always
selected first from which to be recalled. If this recall from the primary copy tape fails or is not
accessible, then IBM Spectrum Archive EE automatically retries the transparent recall
operation against the other replicas if they exist.
Note: Transparent recall is used most frequently because it is activated when you access a
migrated file, such as reading a file.
With multiple tape libraries configured, the eeadm recall command requires the -l option to
specify the tape library from which to recall. When a file is recalled, the recall can occur on
any of the tapes (that is, either primary or redundant copies) from the specified tape library.
The following conditions are applied to determine the best replica:
The condition of the tape
If a tape is mounted
If a tape is mounting
If there are tasks that are assigned to a tape
If conditions are equal between certain tapes, the primary tape is preferred over the
redundant copy tapes. The secondary tape is preferred over the third tape. These rules are
necessary to make the tape selection predictive. However, there are no automatic retries likes
with transparent recalls.
For example, if a primary tape is not mounted but a redundant copy is, the redundant copy
tape is used for the recall task to avoid unnecessary mount operations.
If the specified tape library does not have any replicas, IBM Spectrum Archive EE
automatically resubmits the request to the other tape library to process the bulk recalls:
Three copies: TAPE1@Library1 TAPE2@Library1 TAPE3@Library2
– If -l Library1 → TAPE1 or TAPE2
– If -l Library2 → TAPE3
Two copies: TAPE1@Library1 TAPE2@Library1
– If -l Library1 → TAPE1 or TAPE2
– If -l Library2 → TAPE1 or TAPE2
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 191
The eeadm recall command
The eeadm recall command is used to recall non-empty regular files from tape. The eeadm
recall command uses the following syntax:
eeadm recall <inputfile> [OPTIONS]
The <inputfile> may contain one of two formats. Each line ends with “-- <filename>”
with a space before and after the double dash. Or each line contains a file name with an
absolute path or a relative path that is based on the working directory.The eeadm recall
command with the output of another command.
The eeadm recall command can take as input the output of other commands through a pipe.
In Example 6-77, all files with names ending with .bin are recalled under the /ibm/gpfs/prod
directory, including subdirectories. Therefore, it is convenient to recall whole directories with a
simple command.
Example 6-77 The eeadm recall command with the output of another command
[root@saitama2 prod]# find /ibm/gpfs/prod -name "*.bin" -print | eeadm recall -l lib_saitama
2019-01-09 15:55:02 GLESL277I: The "eeadm recall command" is called without specifying an input file waiting for
standard input.
If necessary press ^D to exit.
2019-01-09 15:55:02 GLESL268I: 4 file name(s) have been provided to recall.
2019-01-09 15:55:03 GLESL700I: Task selective_recall was created successfully, task id is 7112.
2019-01-09 15:55:09 GLESL263I: Recall result: 4 succeeded, 0 failed, 0 duplicate, 0 not migrated, 0 not found, 0
unknown.
6.14.3 Read Starts Recalls: Early trigger for recalling a migrated file
IBM Spectrum Archive EE can define a stub size for migrated files so that the stub size initial
bytes of a migrated file are kept on disk while the entire file is migrated to tape. The migrated
file bytes that are kept on the disk are called the stub. Reading from the stub does not trigger
a recall of the rest of the file. After the file is read beyond the stub, the recall is triggered. The
recall might take a long time while the entire file is read from tape because a tape mount
might be required, and it takes time to position the tape before data can be recalled from tape.
When Read Start Recalls (RSR) is enabled for a file, the first read from the stub file triggers a
recall of the complete file in the background (asynchronous). Reads from the stubs are still
possible while the rest of the file is being recalled. After the rest of the file is recalled to disks,
reads from any file part are possible.
With the Preview Size (PS) value, a preview size can be set to define the initial file part size
for which any reads from the resident file part does not trigger a recall. Typically, the PS value
is large enough to see whether a recall of the rest of the file is required without triggering a
recall for reading from every stub. This process is important to prevent unintended massive
recalls. The PS value can be set only smaller than or equal to the stub size.
This feature is useful, for example, when playing migrated video files. While the initial stub
size part of a video file is played, the rest of the video file can be recalled to prevent a pause
when it plays beyond the stub size. You must set the stub size and preview size to be large
enough to buffer the time that is required to recall the file from tape without triggering recall
storms.
192 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Use the following dsmmigfs command options to set both the stub size and preview size of
the file system being managed by IBM Spectrum Archive EE:
dsmmigfs Update -STUBsize
dsmmigfs Update -PREViewsize
The value for the STUBsize is a multiple of the IBM Spectrum Scale file system’s block size.
this value can be obtained by running the mmlsfs <filesystem>. The PREViewsize parameter
must be equal to or less than the STUBsize value. Both parameters take a positive integer in
bytes.
Example 6-78 shows how to set both the STUBsize and PREViewsize on the IBM Spectrum
Scale file system.
For more information about the dsmmigfs update command, see IBM Documentation.
In a recall, IBM Spectrum Archive EE will queue the tasks to send requests in an order that is
based on tape and starting blocks (refer to 6.14, “Recall” on page 190). This reading
optimization is efficient when significantly large numbers of files that are located in linear
positions of the tape are requested. However, in a case where the requested files are
distributed across multiple bands and wraps, this optimization method was not as effective
since the data could not be read efficiently with a simple linear read.
RAO was introduced to address such cases. Though additional calculation time will be
needed to create the recommended reading order of files, effective use of RAO will decrease
read time significantly in an ideal case.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 193
Figure 6-1 is a conceptual diagram displaying the seeking path of five files based on block
order. The dotted arrows represents the servo head seek path, in this case, starting from File
1 and ending with File 5.
Figure 6-1 A conceptual diagram displaying the seek path of files based on block order
In contrast, Figure 6-2 shows the servo head movement when same files are read based on
RAO. When compared with the previous figure, the total seek path length represented with
the dotted arrow will be significantly shorter in RAO, therefore indicating that the total recall
time will be reduced.
Figure 6-2 A conceptual diagram displaying the seek path of files based on RAO
Though RAO is a powerful feature, there are cases where it will not be effective. As
addressed above, one example of such cases is when there are massive numbers of files to
read from a cartridge. In such cases, a block sorted read of a cartridge is the more
reasonable choice.
To address such cases, IBM Spectrum Archive EE provides an automatic transition of the two
methods during recall requests.
194 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
This automatic transition includes the following features:
File queuing on recall requests to same cartridges
When possible, multiple recall requests to a single cartridge will be joined into a single
RAO request queue. In this case, the RAO requests will be joined up to a 2,700 file queue
in prospect to reduce the total reading time.
RAO usage decision on cases where large number of files are recalled
IBM Spectrum Archive EE will automatically switch to a block order recalling method when
recall requests sum to 5,400 file mark for a single cartridge.2
RAO usage decision on supported cartridges and drives
RAO function will not be used if there are no compatible drives and media available in the
library. Refer to the IBM Documentation - Supported tape drives and media for compatible
drives and media.
Note: For more information on RAO interface specification, refer to the INCITS 503-202x:
Information technology - SCSI Stream Commands - 5 (SSC-5) Overview standard
document.
Leaving the RAO usage on (or in “auto”) is generally recommended in most cases, though
turning the RAO feature off may come to consideration if:
Recalls are always made with large numbers of files (more than 5400 files) from a single
cartridge, at a time.
IBM Spectrum Archive EE is unable to support the RAO feature since there are no
supported drives or media in the library.
RAO function can be globally enabled or disabled with the recall_use_rao option in the
eeadm cluster set command. The default setting is “auto”. Example shows how to disable
and enable the RAO feature.Globally configuring the RAO feature
Current settings can be found with the cluster show command (see Example 6-79).
2 The 5,400 file mark is based on laboratory experimental data. In the experiment, we compared the effect of
recalling N number of 20MB files with the RAO method versus the starting block order method. The total net recall
time indicated that the starting block order method becomes more efficient just over N=5,000 files, and the RAO
performance are significantly less effective by N=10,000 files.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 195
The global RAO usage switch can be overridden locally in a recall command with the
--use-rao option (see Example 6-80).
Example 6-80 Overriding global RAO switch with local recall command
The eeadm recall --resident command will recall the files back initially if the files are
migrated then mark the files resident and remove the link between disk and tape. If the files
are already premigrated the recall operation will be skipped and will just mark the files
resident. This option removes metadata on IBM Spectrum Scale, which is used for keeping
the file/object state.
Note: migreclist.txt contains the list of files to be recalled. In this case, it contains only
‘migrecfile2.txt’.
196 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.16 Reconciliation
This section describes file reconciliation with IBM Spectrum Archive EE and presents
considerations for the reconciliation process.
HSM is not notified upon moves, renames, or deletions of files in IBM Spectrum Scale.
Therefore, over time the metadata of migrated files on IBM Spectrum Scale can diverge from
their equivalents on LTFS. The goal of the reconciliation function is to synchronize the IBM
Spectrum Scale namespace with the corresponding LTFS namespace (per tape cartridge)
and the corresponding LTFS attributes (per tape cartridge).
The reconciliation process resolves any inconsistencies that develop between files in the IBM
Spectrum Scale and their equivalents in IBM Spectrum Archive EE. When files are deleted,
moved, or renamed in IBM Spectrum Scale, the metadata of those files becomes out of sync
with their copies in LTFS.
For each file that was deleted in IBM Spectrum Scale, the reconciliation process deletes the
corresponding LTFS files and symbolic links. If the parent directory of the deleted symbolic
link is empty, the parent directory is also deleted. This process frees capacity resources that
were needed for storing the LTFS index entries of those deleted files.
For each IBM Spectrum Scale file that was moved or renamed in IBM Spectrum Scale, the
reconciliation process updates for each LTFS instance (replica) of that IBM Spectrum Scale
file the LTFS extended attribute that contains the IBM Spectrum Scale path and the LTFS
symbolic link.
Reconciliation can be performed on one or more GPFS file systems, one or more tape
cartridge pools, or a set of tape cartridges. When the reconciliation process involves multiple
tape cartridges, multiple IBM Spectrum Archive EE nodes and tape drives can be used in
parallel. However, because recall tasks have priority, only available tape drives are used for
reconciliation. After reconciliation is started, the tape cartridge cannot be unmounted until the
process completes.
The synchronization and update to the IBM Spectrum Archive EE tapes can be time
consuming and normally only required before any eeadm tape export command. Normally if
files are deleted from the IBM Spectrum Scale file systems, merely the update for the amount
of reclaimable space is sufficient information. In other words, as files are deleted,
administrators only want to know how much space can be reclaimed. Thus starting with
v1.3.0.0, the default for eeadm tape reconcile is to only update internal metadata information
in order to correctly reflect the Reclaimable% column of the eeadm tape list command. In
order to update the tape (like previous to v1.3.0.0), the --commit-to-tape option should be
supplied as part of the eeadm tape reconcile command.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 197
2. After a reconciliation process is started, new migration tasks are prevented until the
reconciliation process completes on the reconciling tapes. However, if any migration tasks
are running, the reconciliation process does not begin until all migration tasks complete.
3. Recalls from a tape cartridge being reconciled are not available while the reconciliation
process is updating the index for that tape cartridge, which is a short step in the overall
reconciliation process.
The command outputs in the following examples show the effect that reconciliation has on the
Reclaimable% column after that file is deleted from the IBM Spectrum Scale file system.
Example 6-82 shows the initial file state of a single file on tape.
The file is also present on the IBM Spectrum Scale file system, as shown in Example 6-83.
Example 6-83 List the file on the IBM Spectrum Scale file system
[root@ueno ~]# ls -hl
/gpfs/gpfs0/cmt/md1/files/LTFS_EE_FILE_mQH4oCZXQZKSXc6cYZjAuXPnAFq56QBWJmY66PcuVnA
IW6BkrK_gzCwGJ.bin
-rw------- 1 root root 1.8G Nov 12 15:29
/gpfs/gpfs0/cmt/md1/files/LTFS_EE_FILE_mQH4oCZXQZKSXc6cYZjAuXPnAFq56QBWJmY66PcuVnA
IW6BkrK_gzCwGJ.bin
198 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
At this point, this file can be removed with the rm command. After the file is removed, IBM
Spectrum Archive EE has the reclaimable space information from eeadm tape list for these
tapes (see Example 6-84).
If you perform a reconciliation of the tape now using the default settings, IBM Spectrum
Archive EE only updates the amount of reclaimable space on the tape within the internal
metadata structure, as shown in Example 6-85.
If you now list the tapes with eeadm tape list, you may see that the reclaimable space
percentage be increased due to the deleted files (because it is in percentage and not in GBs,
the size of deletion need to be large to see a percentage change).
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 199
6.17 Reclamation
The space on tape that is occupied by deleted files is not reused during normal IBM Spectrum
Archive EE operations. New data is always written after the last index on tape. The process of
reclamation is similar to the same named process in IBM Spectrum Protect environment
because all active files are consolidated onto a new, empty, second tape cartridge. This
process improves overall tape usage and utilization.
When files are deleted, overwritten, or edited on IBM Spectrum Archive EE tape cartridges, it
is possible to reclaim the space. The reclamation function of IBM Spectrum Archive EE frees
tape space that is occupied by non-referenced files and non-referenced content that is
present on the tape. The reclamation process copies the files that are referenced by the LTFS
index of the tape cartridge being reclaimed to another tape cartridge, updates the GPFS/IBM
Spectrum Scale inode information, and then reformats the tape cartridge that is being
reclaimed.
Use the eeadm tape reclaim command to start reclamation of a specified tape cartridge pool
or of certain tape cartridges within a specified tape cartridge pool. The eeadm tape reclaim
command is also used to specify thresholds that indicate when reclamation is performed by
the percentage of the available capacity on a tape cartridge.
Example 6-86 shows the results of reclaiming a single tape cartridge 058AGWL5.
200 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
2021-03-24 04:08:49 GLESR216I: Multiple processes started in parallel. The maximum
number of processes is unlimited.
2021-03-24 04:08:49 GLESL084I: Start reclaiming 1 tapes in the following list of
tapes:
2021-03-24 04:08:49 GLESR212I: Source candidates: JCB370JC .
2021-03-24 04:08:49 GLESL677I: Files on tape JCB370JC will be copied to tape
JCA561JC.
2021-03-24 04:14:30 GLESL085I: Tape JCB370JC successfully reclaimed, it remains in
tape pool Arnoldpool2.
2021-03-24 04:14:30 GLESL080I: Reclamation complete. 1 tapes reclaimed, 0 tapes
unassigned from the tape pool.
At the end of the process, the tape cartridge is reformatted. The tape remains in the tape
cartridge pool unless the “--unassign” option is specified. For more information, see , “The
eeadm <resource type> --help command” on page 316.
Example 6-87 shows a tape in the check_key_server state and using the eeadm tape
validate to restore its state back to appendable or full after fixing the encryption key server.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 201
Tape ID Status State Usable(GiB) Used(GiB) Available(GiB)
Reclaimable% Pool Library Location Task ID
MB0021JM ok appendable 4536 4536 0 0% pool2
lib_mikasa homeslot -
Two other tape states exist that require the user to replace the tapes. These two states occur
when a tape encounters a read or write failure, and can be remedied by using the eeadm tape
replace command. This command moves the data off the bad tape and onto an appendable
tape within the same pool maintaining the migration order and finally removing the bad tape
from the pool. The following are the two replacement tape states:
need_replace
The IBM Spectrum Archive EE system detected one or more permanent read errors on
this tape
require_replace
The IBM Spectrum Archive EE system detected a permanent write error on the tape.
Example 6-88 shows the output of running the eeadm tape replace command
For more information about the various tape cartridge states, see 10.1.4, “Tape status codes”
on page 320.
6.19.1 Importing
Import tape cartridges to your IBM Spectrum Archive EE system by running the eeadm
tape import <list_of_tapes> -p <pool> [-l <library>] [OPTIONS] command.
When you import a tape cartridge, the eeadm tape import command performs the following
actions:
1. Adds the specified tape cartridge to the IBM Spectrum Archive EE library
2. Assigns the tape to the designated pool
3. Adds the file stubs in an import directory within the IBM Spectrum Scale file system
202 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 6-89 shows the import of an LTFS tape cartridge that was created on a different
LTFS system into a directory that is called FC0257L8 in the /ibm/gpfs file system.
Example 6-91 on page 203 shows the use of the -P parameter, which can be used to redirect
the imported files to an alternative directory. The VOLSER is still used in the directory name,
but you can now specify a custom import file path by using the -P option. If the specified path
does not exist, it is created.
With each of these parameters, you have the option of renaming imported files by using the
--rename parameters. This option will only rename any imported files with conflicting names
by appending the suffix “_i” where i is a number from 1 to n.
When the tape cartridge is offline and outside the library, the IBM Spectrum Scale offline files
on disk or the files on tape cartridge should not be modified.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 203
Example 6-92 shows an example of making an offline tape cartridge online.
The export process then removes all files from the IBM Spectrum Scale file system that exist
on the exported tape cartridge. The files on the tape cartridges are unchanged by the export,
and are accessible by other LTFS systems.
Export considerations
Consider the following information when planning IBM Spectrum Archive EE export activities:
If you put different logical parts of an IBM Spectrum Scale namespace (such as the project
directory) into different LTFS tape cartridge pools, you can export tape cartridges that
contain the entire IBM Spectrum Scale namespace or only the files from a specific
directory within the namespace.
Otherwise, you must first recall all the files from the namespace of interest (such as the
project directory), then migrate the recalled files to an empty tape cartridge pool, and then
export that tape cartridge pool.
Reconcile occurs automatically before the export is processed.
Although the practice is not preferable, tape cartridges can be physically removed from IBM
Spectrum Archive EE without exporting them. In this case, no changes are made to the IBM
Spectrum Scale inode. The following results can occur:
Causes a file operation that requires access to the removed tape cartridge to fail. No
information as to where the tape cartridges are is available.
Files on an LTFS tape cartridge can be replaced in IBM Spectrum Archive EE without
reimporting (that is, without updating anything in IBM Spectrum Scale). This process is
equivalent to a library going offline and then being brought back online without taking any
action in the IBM Spectrum Scale namespace or management.
Important: If a tape cartridge is removed from the library without the use of the export
utility, modified, and then reinserted in the library, the behavior can be unpredictable.
204 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 6-93 shows the typical output from the export command.
Example 6-94 on page 205 shows how the normal exported tape is displayed as exported by
running the eeadm info tapes command.
Tape ID Status State Usable(GiB) Used(GiB) Available(GiB) Reclaimable% Pool Library Location Task ID
FC0260L8 ok appendable 10907 7 10899 0% temp liba homeslot -
UEF108M8 ok appendable 5338 46 5292 0% pool4 liba homeslot -
DV1993L7 ok appendable 5338 37 5301 0% pool4 liba homeslot -
FC0255L8 ok exported 0 0 0 0% - liba homeslot -
If errors occur during the export phase, the tape goes to the export state. However, some of
the files that belong to that tape might still remain in the file system and still have a reference
to that tape. Such an error can occur when an export is happening and while reconciliation
occurs one starts to modify the files belonging to the exporting tape. In such a scenario, see
9.5, “Software” on page 303 on how to clean up the remaining files on the IBM Spectrum
Scale file system.
Regarding full replica support, Export/Import does not depend on the primary/redundant
copy. When all copies are exported, the file is exported.
Table 6-1 lists a use case example where a file was migrated to three physical tapes: TAPE1,
TAPE2, TAPE3. The file behaves as shown by export operations.
Table 6-1 Export operations use case scenario of file with three tapes
Operation File
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 205
6.19.3 Offlining tape cartridges
Offline tape cartridges from your IBM Spectrum Archive EE system by running the eeadm
tape offline command. This process marks all files from the tape cartridges or tape
cartridge pool that are specified and marks them offline and those files cannot be accessed.
However, the corresponding inode of each file is kept in IBM Spectrum Scale. Those files can
be brought back to the IBM Spectrum Scale namespace by making online the tape cartridge
by using the eeadm tape online command.
If you want to move tape cartridges to an off-site location for DR purposes but still retain files
in the IBM Spectrum Scale file system, follow the procedure that is described here. In
Example 6-95, tape JCB610JC contains redundant copies of five MPEG files that must be
moved off-site.
If you run the eeadm tape list command, you can see the offline status of the tape cartridge,
as shown in Example 6-96.
It is now possible to physically remove tape JCA561JC from the tape library so that it can be
sent to the off-site storage location.
Regarding full replica support, Offlining/Onlining does not depend on the primary/redundant
copy. When all copies are offlined, the file is offline.
Table 6-2 lists a use case example where a file was migrated to three physical tapes: TAPE1,
TAPE2, TAPE3. The file behaves as shown by offline operations.
Table 6-2 Export operations use case scenario of file with three tapes
Operation File
TAPE1/TAPE2 are offline exported, and then File is offline (cannot recall, and the file will be
TAPE3 is exported. only recalled from TAPE1/TAPE2 once they are
online).
206 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 6-97 shows the file state when the tape host has been offlined.
Table 6-3 lists the available IBM Spectrum Archive EE drive attributes for the attached
physical tape drives.
Migration If the Migration attribute is set for a drive, that drive can process migration tasks.
If not, IBM Spectrum Archive EE never runs migration tasks by using that drive.
Save tasks are also allowed/disallowed through this attribute setting. It is
preferable that there be at least one tape drive that has this attribute set to
Migration.
Recall If the Recall attribute is set for a drive, that drive can process recall tasks. If not,
IBM Spectrum Archive EE never runs recall tasks by using that drive. Both
automatic file recall and selective file recall are enabled/disabled by using this
single attribute. There is no way to enable/disable one of these two recall types
selectively. It is preferable that there be at least one tape drive that has this
attribute set to Recall.
Generic If the Generic attribute is set for a drive, that drive can process generic tasks. If
not, IBM Spectrum Archive EE never runs generic tasks by using that drive. IBM
Spectrum Archive EE creates and runs miscellaneous generic tasks for
administrative purposes, such as formatting tape, checking tape, reconciling
tape, reclaiming a tape, and validating a tape. Some of those tasks are internally
run with any of the user operations. It is preferable that there be at least one tape
drive that has this attribute set to Generic. For reclaiming tape, at least two tape
drives are required, so at least two drives need the Generic attribute.
To set these attributes for a tape drive, the attributes can be specified when adding a tape
drive to IBM Spectrum Archive EE. Use the following command syntax:
eeadm drive assign <drive serial> -n <node_id> -r
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 207
The -r option requires a decimal numeric parameter. A logical OR applies to set the three
attributes: Migrate (4), Recall (2), and Generic (1). For example, a number of 6 for -r
allows migration and recall task while copy and generic task are disallowed. All of the
attributes are set by default.
To check the current active drive attributes, the eeadm drive list command is useful. This
command shows each tape drive’s attributes, as shown in Example 6-98.
The letters m, r, and g are shown when the corresponding attribute Migration, Recall, and
Generic are set to on. If an attribute is not set, “-” is shown instead.
The role of an assigned drive can be modified by using the eeadm drive set command. For
example, the following command changes the role to Migrate and Recall.
eeadm drive set <drive serial> -a role -v mr
The configuration change takes effect after the on-going process has completed.
Drive attributes setting hint: In a multiple nodes environment, it is expected that the
reclaim driver works faster if two tape drives that are used for reclaim are assigned to a
single node. For that purpose, tape drives with the Generic attribute should be assigned to
a single node and all of other drives of the remaining nodes should not have the Generic
attribute.
Note: An intermix of LTO and TS11xx tape drive technology and media on a single library
is not supported by IBM Spectrum Archive EE.
208 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The following main use cases are expected to be used by this feature:
Tape media and technology migration (from old to new generation tapes)
Continue using prior generation formatted tapes (read or write) with the current technology
tape drive generation
To generate and use a mixed tape drive environment, you must define the different LTO or
TS11xx drive types with the creation of the logical library partition (within your tape library) to
be used along with your IBM Spectrum Archive EE setup.
When LTO-9, 8, M8, 7, 6, and 5 tapes are used in a tape library, correct cartridges and drives
are selected by IBM Spectrum Archive EE to read or write the required data. For more
information, see this IBM Documentation web page.
When 3592 JC, JD, and JE tapes are used in a tape library and IBM TS1160, TS1155,
TS1150, and TS1140 drives are used correct tapes and drives are selected by IBM Spectrum
Archive EE to read or write the required data.
With this function, a data migration between different generations of tape cartridges can be
achieved. You can select and configure which TS11xx format (TS1155, TS1150, or TS1140)
is used by IBM Spectrum Archive EE for operating 3592 JC tapes. The default for IBM
Spectrum Archive EE is always to use and format to the highest available capacity. The
TS1160 supports only recalls of JC media formatted using TS1140.
The eeadm tape assign command can be used for pool configuration when new physical
tape cartridges are added to your IBM Spectrum Archive EE setup:
eeadm tape assign <list_of_tapes> -p <pool> [-l <library>] [OPTIONS]
WORM support for the IBM TS1160, TS1155, TS1150, and TS1140 tape drives
From the long-term archive perspective, there is sometimes a requirement to store files
without any modification that is ensured by the system. You can deploy Write Once Read
Many (WORM) tape cartridges in your IBM Spectrum Archive EE setup. Only 3592 WORM
tapes that can be used with IBM TS1160, TS1155, TS1150, or TS1140 drives are supported.
Note: LTO WORM tapes are not supported for IBM Spectrum Archive EE.
For more information about IBM tape media and WORM tapes, see this website.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 209
Does not ensure that the file cannot be modified through GPFS in the following ways:
– Migrate the immutable files to tape.
– Recall the immutable files to disk.
– Change the immutable attribute of the file on disk and modify.
Example 6-99 shows how to set the WORM attribute to an IBM Spectrum Archive EE pool by
using the eeadm pool create command.
Example 6-99 Set the WORM attribute to an IBM Spectrum Archive EE pool
[root@saitama2 prod]# eeadm pool create myWORM --worm physical
There is also an IBM Spectrum Scale layer that can provide a certain immutability for files
within the GPFS file system. You can apply immutable and appendOnly restrictions either to
individual files within a file set or to a directory. An immutable file cannot be changed or
renamed. An appendOnly file allows append operations, but not delete, modify, or rename
operations.
An immutable directory cannot be deleted or renamed, and files cannot be added or deleted
under such a directory. An appendOnly directory allows new files or subdirectories to be
created with 0-byte length. All such new created files and subdirectories are marked as
appendOnly automatically.
The immutable flag and the appendOnly flag can be set independently. If both immutability
and appendOnly are set on a file, immutability restrictions are in effect.
To set or unset these attributes, use the following IBM Spectrum Scale command options:
mmchattr -i yes|no
This command sets or unsets a file to or from an immutable state:
– -i yes
Sets the immutable attribute of the file to yes.
– -i no
Sets the immutable attribute of the file to no.
mmchattr -a yes|no
This command sets or unsets a file to or from an appendOnly state:
– -a yes
Sets the appendOnly attribute of the file to yes.
– -a no
Sets the appendOnly attribute of the file to no.
Note: Before an immutable or appendOnly file can be deleted, you must change it to
mutable or set appendOnly to no (by using the mmchattr command).
210 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Storage pool assignment of an immutable or appendOnly file can be changed. An immutable or
appendOnly file is allowed to transfer from one storage pool to another.
The system displays information similar to the output that is shown in Example 6-100.
Note: The immutable and appendOnly flag check takes effect after the file is closed.
Therefore, the file can be modified if it is opened before the file is changed to
immutable.
mode
An immutable or appendOnly file’s mode cannot be changed.
ownership, acl
These attributes cannot be changed for an immutable or appendOnly file.
timestamp
The time stamp of an immutable or appendOnly file can be changed.
directory
If a directory is marked as immutable, no files can be created, renamed, or deleted under
that directory. However, a subdirectory under an immutable directory remains mutable
unless it is explicitly changed by the mmchattr command.
If a directory is marked as appendOnly, no files can be renamed or deleted under that
directory. However, 0-byte length files can be created.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 211
For more information about IBM Spectrum Scale V5.1.2 immutable and appendOnly
limitations, see IBM Documentation.
Example 6-101 shows the output that you receive while working, showing, and changing the
IBM Spectrum Scale immutable or appendOnly file attributes.
Example 6-101 Set or change an IBM Spectrum Scale file immutable file attribute
[root@ltfsee_node0]# echo "Jan" > jan_jonas.out
[root@ltfsee_node0]# mmlsattr -L -d jan_jonas.out
file name: jan_jonas.out
metadata replication: 1 max 2
data replication: 1 max 2
immutable: no
appendOnly: no
flags:
storage pool name: system
fileset name: root
snapshot name:
creation time: Mon Aug 31 15:40:54 2015
Windows attributes: ARCHIVE
Encrypted: yes
gpfs.Encryption:
0x454147430001008C525B9D470000000000010001000200200008000254E60BA4024AC1D500010001
00010003000300012008921539C65F5614BA58F71FC97A46771B9195846A9A90F394DE67C4B9052052
303A82494546897FA229074B45592D61363532323261642D653862632D346663632D383961332D3461
37633534643431383163004D495A554E4F00
EncPar 'AES:256:XTS:FEK:HMACSHA512'
type: wrapped FEK WrpPar 'AES:KWRAP' CmbPar 'XORHMACSHA512'
KEY-a65222ad-e8bc-4fcc-89a3-4a7c54d4181c:ltfssn2
212 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
-bash: jan_jonas.out: Read-only file system
[root@ltfsee_node0]#
These immutable or appendOnly file attributes can be changed at any time by the IBM
Spectrum Scale administrator, so IBM Spectrum Scale is not able to provide an complete
immutability.
If you are working with IBM Spectrum Archive EE and IBM Spectrum Scale and you plan
implementing a WORM solution along with WORM tape cartridges, these two main
assumptions apply:
Only files that have IBM Spectrum Scale with the immutable attribute ensure no
modification.
The IBM Spectrum Scale immutable attribute is not changed after it is set unless it is
changed by an administrator.
Consider the following limitations when using WORM tapes together with IBM Spectrum
Archive EE:
WORM tapes are supported only with IBM TS1160, TS1155, TS1150, and TS1140 tape
drives (3592 JV, JY, JZ).
If IBM Spectrum Scale immutable attributes are changed to yes after migration, the next
migration fails against the same WORM pool.
IBM Spectrum Archive EE supports the following operations with WORM media:
– Migrate
– Recall
– Offline export and offline import
IBM Spectrum Archive EE does not support the following operations with WORM media:
– Reclaim
– Reconcile
– Export and Import
For more information about the IBM Spectrum Archive EE commands, see 10.1,
“Command-line reference” on page 316.
You can use the eeadm file state command to discover the physical location of files. To help
with the management of replicas, this command also indicates which tape cartridges are
used by a particular file, how many replicas exist, and the health state of the tape.
Example 6-102 shows the typical output of the eeadm file state command. Some files are
on multiple tape cartridges, some are in a migrated state, and others are premigrated only.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 213
Replicas: 2
Tape 1: JCB610JC@test3@lib_saitama (tape state=appendable)
Tape 2: JD0321JD@test4@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/file2.bin
State: migrated
ID: 11151648183451819981-3451383879228984073-2015134857-974348-0
Replicas: 2
Tape 1: JCB610JC@test3@lib_saitama (tape state=appendable)
Tape 2: JD0321JD@test4@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/file3.bin
State: migrated
ID: 11151648183451819981-3451383879228984073-599546382-974350-0
Replicas: 2
Tape 1: JCB610JC@test3@lib_saitama (tape state=appendable)
Tape 2: JD0321JD@test4@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/file4.bin
State: migrated
ID: 11151648183451819981-3451383879228984073-2104982795-3068894-0
Replicas: 1
Tape 1: JD0321JD@test4@lib_saitama (tape state=appendable)
For more information about supported characters for file names and directory path names,
see IBM Documentation.
Example 6-103 shows the command that is used to display all IBM Spectrum Archive EE tape
cartridge pools.
214 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
test2 6.1 0.0 6.1 0% 1 3592
lib_saitama G0
Example 6-104 shows the serial numbers and status of the tape drives that are used by IBM
Spectrum Archive EE.
To view all the IBM Spectrum Archive EE tape cartridges, run the command that is shown in
Example 6-105.
Regularly monitor the output of IBM Spectrum Archive EE tasks to ensure that tasks are
progressing as expected by using the eeadm task list and eeadm task show command.
Example 6-106 shows a list of active tasks.
The eeadm node list command (see Example 6-107) provides a summary of the state of
each IBM Spectrum Archive EE component node.
The eeadm task show command (see Example 6-108) provides a summary of the specified
active task in IBM Spectrum Archive EE. The task id corresponds to the task id that is
reported by the eeadm task list command.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 215
=== Task Information ===
Task ID: 7168
Task Type: selective_recall
Command Parameters: eeadm recall mig -l lib_saitama
Status: running
Result: -
Accepted Time: Thu Jan 10 16:27:30 2019 (-0700)
Started Time: Thu Jan 10 16:27:30 2019 (-0700)
Completed Time: -
In-use Libraries: lib_saitama
In-use Node Groups: G0
In-use Pools: test3
In-use Tape Drives: 0000078PG20E
In-use Tapes: JCB610JC
Workload: 3 files, 5356750 bytes in total to recall in this task.
Progress: 1 completed (or failed) files / 3 total files.
Result Summary: -
Messages:
2019-01-10 16:27:30.421251 GLESM332W: File /ibm/gpfs/prod/LTFS_EE_FILE_2dEPRHhh_M.bin is not migrated.
Table 6-4 lists SNMP traps that can be issued, showing the error message that is generated,
along with the trap name, severity code, and the OID for each trap.
216 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Description Name Severity OID
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 217
Description Name Severity OID
To configure Net-SNMP, complete the following steps on each IBM Spectrum Archive EE
node:
1. Open the /etc/snmp/snmpd.conf configuration file.
2. Add the following entry to the file:
master agentx
trap2sink <managementhost>
The variable <managementhost> is the host name or IP address of the host to which the
SNMP traps are sent.
3. Open the /etc/snmp/snmptrapd.conf configuration file.
4. Add the following entry to the file:
disableauthorization yes
5. Restart the SNMP daemon by running the following command:
[root@ltfs97 ~]# systemctl restart snmpd.service
218 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.25.1 Starting and stopping the snmpd daemon
Before IBM Spectrum Archive EE is started, you must start the snmpd daemon on all nodes
where IBM Spectrum Archive EE is running.
Example 6-109 SNMP trap example of an IBM Spectrum Archive EE node that has a low pool threshold
2018-11-26 09:08:42 tora.tuc.stglabs.ibm.com [UDP: [9.11.244.63]:60811->[9.11.244.63]:162]:
DISMAN-EVENT-MIB::sysUpTimeInstance = Timeticks: (147206568) 17 days, 0:54:25.68
SNMPv2-MIB::snmpTrapOID.0 = OID: IBMSA-MIB::ibmsaWarnM609PoolLowSpace
IBMSA-MIB::ibmsaMessageSeverity.0 = INTEGER: warning(40) IBMSA-MIB::ibmsaJob.0 = INTEGER:
other(7) IBMSA-MIB::ibmsaEventNode.0 = STRING: "tora.tuc.stglabs.ibm.com"
IBMSA-MIB::ibmsaMessageText.0 = STRING: "GLESM609W: Pool space is going to be small, library:
lib_tora, pool: je_pool1, available capacity: 23.6(TiB), threshold:
30(TiB)"
The IBM Spectrum Archive EE REST API can be accessed in two ways. The first is through a
terminal window with the curl command and the second way is through a web browser. Both
ways output the same data. In this section, the curl command is used.
The following are supported GET operations when executing rest commands:
pretty
Specify for pretty-printing. The default value is false.
sort: <string>(,<string>...)
Specify field name or names to use as sort key. The default sort order is ascending. Use
the “-” sign to sort in descending order.
fields: <string>(,<string>...)
Specify field names that are to be included in the response.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 219
Note: The examples in this section are all performed on the server with the REST RPM
installed and uses localhost to request resources. When accessing the REST API over a
remote server, replace localhost with the server ip with the rest rpms installed.
The following is an example command of calling the pools endpoint using localhost:
curl -X GET ‘http://localhost:7100/ibmsa/v1/pools/’
The following is the response data returned when requesting for the pools endpoint:
id: <string>
UUID of Pool, assigned by system at the creation of pool.
name: <string>
User-specified name of pool.
capacity: <number>
Total capacity of tapes assigned to the pool, in bytes. The capacity = used_space +
free_space.
mode: <string>
The current operational mode of the pool. Access to the member tapes is temporarily
disabled when this field is set to “disabled”, “relocation_source”, or
“relocation_destination”. under normal operating conditions, the field is set to “normal”. If
an internal error occurs, the field is set to an empty string.
used_space: <number>
Used space of the pool, in bytes. The used_space = active_space + reclaimable_space.
free_space: <number>
Free space of the pool, in bytes.
active_space: <number>
Active space (used space consumed by active-referred files) of the pool, in bytes.
reclaimable_space: <number>
The reclaimable space (used space consumed by unreferred files) of the pool, in bytes.
Note that this is the amount of estimated size of the unreferenced space that is available
for reclamation on the assigned tapes.
non_appendable_space: <number>
The total capacity, in bytes, of tapes in the pool that cannot be written to in the format
specified for the pool, and that don’t match the media_restriction value for the pool. The
format and the media_restriction values are provided as attributes of the pool.
num_of_tapes: <number>
Number of tapes assigned to the pool.
format_class: <string>
The format class of the pool.
220 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
media_restriction: <string>
The media_restriction setting is stored in the regular expression of the bar code value.
The media_restriction is used to define the type of cartridge that can be used for writing.
The cartridge media type that is represented by the last two letters of the cartridge bar
code is used in this field. The string can be either “^.{6}XX$”, “^.{8}$”, or “unknown”. The
“^.{6}XX$” represents any 6 characters followed by type “XX “, where “XX “ is one of the
following cartridge media types: L5, L6, L7, L8, M8, L9, JB, JC, JD, JE, JK, JL, JY, or JZ.
The “^.{8}$” represents any 8 characters, and means that any cartridge media type is
acceptable. A value of “unknown” means that there is an error condition.
device_type: <string>
Tape device type that can be added to the pool. Can be either 3592, LTO, or left blank.
worm: <string>
WORM type. Can be either physical, no, or unknown.
fill_policy: <string>
Tape fill policy.
owner: <string>
Owner.
mount_limit: <number>
Maximum number of drives that can be used for migration. 0 means unlimited.
low_space_warning_enable: <bool>
Whether monitoring thread sends SNMP trap for low space pool.
low_space_warning_threshold: <number>
SNMP notification threshold value for free pool size in bytes. 0 when no threshold is set.
no_space_warning_enable: <bool>
Whether monitoring thread sends SNMP trap for no space pool.
library_name: <string>
Library name to which the pool belongs.
library_id: <string>
Library ID (serial number) to which the pool belongs.
node_group: <string>
Node group name to which the pool belongs.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 221
Example 6-110 shows how to request the pools resource through curl commands.
222 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
"worm": "no"
}
]
Example 6-111 shows how to call pools with specifying specific fields and sorting the output
in descending order.
The following is the response data when requesting the tapes endpoint:
id: <string>
Tape ID. Because barcode is unique within a tape library only, the id is in format of
<barcode>@<library_id>.
barcode: <string>
Barcode of the tape.
state: <string>
The string indicates the state of the tape. The string can be either “appendable”,
“append_fenced”, “offline”, “recall_only”, “unassigned”, “exported”, “full”, “data_full”,
“check_tape_library”, “need_replace”, “require_replace”, “require_validate”,
“check_key_server”, “check_hba”, “inaccessible”, “non_supported”, “duplicated”,
“missing”, “disconnected”, “unformatted”, “label_mismatch”, “need_unlock”, and
“unexpected_cond”. If the string is “unexpected_cond”, it is probably an error occurred.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 223
status: <string>
The string indicates the severity level of the tape's status. The string can be either “error”,
“degraded”, “warning”, “info”, or “ok”.
media_type: <string>
Media type of a tape. Media type is set even if the tape is not assigned to any pool yet.
Empty string if the tape is not supported by IBM Spectrum Archive.
media_generation: <string>
Media generation of a tape. Media generation determines a possible format that the tape
can be written in.
format_density: <string>
Format of a tape. Empty string if the tape is not assigned to any pool.
worm: <bool>
Whether WORM is enabled for the tape.
capacity: <number>
Capacity of the tape, in bytes. capacity = used_space + free_space.
Appendable: <string>
A tape that can be written in the format that is specified by the pool attributes, and on the
cartridge media type that is specified by the pool attributes, is appendable. The format and
the cartridge media type are provided as attributes of the pool to which the tape belongs.
The string can be either “yes”, “no”, or it can be empty. If the tape falls into a state such as
“append_fenced” or “inaccessible”, the string becomes “no”. If the string is empty, the tape
is not assigned to any pool.
used_space: <number>
Used space of the tape, in bytes. used_space = active_space + reclaimable_space.
free_space: <number>
Free space of the tape, in bytes.
active_space: <number>
Active space (used space consumed by active-referred files) of the tape, in bytes.
reclaimable_space: <number>
Reclaimable space (used space consumed by unreferred files) of the tape, in bytes. This
amount is the estimated size of the unreferenced space that is available for reclamation on
the tape.
address: <number>
Address of this tape.
drive_id: <string>
Drive serial number that the tape is mounted on. Empty string if the tape is not mounted.
offline_msg: <string>
Offline message that can be specified when performing tape offline.
task_id: <string>
The task_id of the task which using this tape.
224 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
location_type: <string>
The location type of the tape. The string can be either “robot”, “homeslot”, “ieslot”, “drive”,
or empty. If the string is empty, the tape is missing.
library_id: <string>
Library ID (serial number) to which the pool is belongs.
library_name: <string>
Library name that the pool is belongs to
pool_id: <string>
Pool ID to which this tape is assigned. Empty string if the tape is not assigned to any pool.
The following is the response data that is returned when requesting this endpoint:
id: <string>
Serial number of the library.
name: <string>
User-defined name of the library.
model: <string>
Mode type of the library.
serial: <string>
The serial number of the library.
scsi_vendor_id: <string>
The vendor ID of the library.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 225
scsi_firmware_revision: <string>
The firmware revision of the library.
num_of_drives: <number>
The number of tape drives that are assigned to the logical library.
num_of_ieslots: <number>
The number of I/E slots that are assigned to the logical library.
num_of_slots: <number>
The number of storage slots that are assigned to the logical library.
host_device_name: <string>
The Linux device name of the library.
host_scsi_address: <string>
The current data path to the tape drive from the assigned node, in the decimal notation of
host.bus.target.lun.
errors: <array>
An array of strings that represents errors. If no error is found, an empty array is returned.
The available filtering parameter for the libraries endpoint is the name of the library.
The following is the response data that is returned when requesting this endpoint:
id: <string>
Nodegroup ID. Because nodegroup name is unique within a tape library only, the ID is in
format of <nodegroup_name>@<library_id>.
name: <string>
User-specified name of the node group.
num_of_nodes: <number>
The number of nodes assigned to the node group.
library_id: <string>
The library ID (serial number) to which the node group belongs.
library_name: <string>
The name of the library to which the node group belongs.
The available filtering parameters for nodegroups are name (nodegroup name), and
library_name. These parameters filter out nodegroups that do not meet the values that were
passed in.
226 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.26.5 Nodes endpoint
The nodes endpoint returns information about each EE node assigned to the cluster. The
following is an example of calling the nodes endpoint:
curl -X GET ‘http://localhost:7100/ibmsa/v1/nodes/’
The following is the response data when requesting the nodes endpoint:
id: <number>
The node ID, which is the same value as the corresponding value in the IBM Spectrum
Scale node IDs.
ip: <string>
The IP address of the node. Specifically, this is the ‘Primary network IP address’ in GPFS.
hostname: <string>
The host name of the node (the ‘GPFS daemon node interface name’ in GPFS).
port: <number>
The port number for LTFS.
state: <string>
The LE status of the node.
num_of_drives: <number>
The number of drives attached to the node.
control_node: <bool>
True if the node is configured as control node.
active_control_node: <bool>
True if the node is configured as a control node and is active.
enabled: <bool>
True, if the node is enabled
library_id: <string>
The library ID of the library to which the node is attached.
library_name: <string>
The name of the library to which the node is attached.
nodegroup_id: <string>
The ID of the nodegroup to which the node is belongs.
nodegroup_name: <string>
The name of the nodegroup to which the node is belongs.
The available filtering parameters for the nodes endpoint are library_name and
nodegroup_name. These parameters will filter out nodes that do not match the passed in
values.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 227
6.26.6 Drives endpoint
The drives endpoint returns information about each visible drive within the EE cluster. The
following is an example of calling the drives endpoint:
curl -X GET ‘http://localhost:7100/ibmsa/v1/drives/’
The following is the response data when requesting the drives endpoint:
id: <string>
Serial number of the drive.
state: <string>
The drive state. For more information, see Table 10-1, “Status and state codes for
eeadm drive list” on page 317.
status: <string>
The string indicates the severity level of the tape's status.
type: <string>
Drive type, which can be empty if a drive is not assigned to any node group.
role: <string>
Three character string to represent the drive role. Can be empty if a drive is not assigned
to any node group.
address: <number>
The address of the drive within the library.
tape_barcode: <string>
The barcode of a tape to which the drive is mounted. Empty if no tape is mounted on the
drive.
task_id: <string>
The task_id of the task which is using this drive.
library_id: <string>
The ID of the library to which the drive belongs.
library_name: <string>
The name of the library to which the drive belongs.
nodegroup_name: <string>
The name of a nodegroup to which the drive is assigned.
node_id: <string>
The ID of the node to which the drive is assigned.
node_hostname: <string>
The host name of the node to which the drive is assigned. This field can be empty if a
drive is not assigned to any node.
scsi_vendor_id: <string>
The vendor ID of the drive.
scsi_product_id: <string>
The product ID of the drive.
228 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
scsi_firmware_revision: <string>
The firmware revision of the drive.
host_device_name: <string>
The linux device name of the drive.
host_scsi_address: <string>
The current data path to the tape drive from the assigned node, in the decimal notation of
host.bus.target.lun.
The available filtering parameters for the drive endpoint are library_name and
nodegroup_name. These parameters will filter out drives that do not match the passed drives in
values.
The following is the response data when requesting the tasks endpoint:
id: <string>
The task ID. Because the task id is overwritten after the task id reaches the upper limit, the
id is in the format of <task_id>@<created_time>.
task_id: <number>
The same value as the corresponding value in the IBM Spectrum Archive task ID.
type: <string>
The task type which meets the eeadm commands. A value of “unknown” means that there
is an error condition.
cm_param: <string>
The eeadm command that user input. If the type is “transparent_recall”, the field is empty.
result: <string>
The result for completed tasks. The string can be either “succeeded”, “failed”, “aborted”,
“canceled”, “suspended” or it can be empty. If the string is empty, the task is not
completed.
status: <string>
The status for the task. The string can be either “waiting”, “running”, “interrupted”,
“suspending”, “canceling”, “completed” or it can be empty. If the string is empty, an internal
error occurs.
inuse_libs: <string array>
The libraries that the task is currently using. The value is a string array of library.name.
inuse_node_groups: <string array>
The node groups that the task is currently using. The value is a string array of
nodegroup.name.
inuse_drives: <string array>
The drives that the task is currently using. The value is a string array of drive.id.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 229
inuse_pools: <string array>
The pools that the task is currently using. The value is a string array of pool.name.
inuse_tapes: <string array>
The tapes that the task is currently using. The value is a string array of tape.barcode.
node_hostname: <string>
The host name of the node to which the drive is assigned. This field can be empty if a
drive is not assigned to any node.
230 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
6.27 File system migration
When refreshing the disk storage and/or IBM Spectrum Scale nodes, the IBM Spectrum
Scale file system may need to be migrated to a new file system. The file system migration can
be within the IBM Spectrum Scale cluster or between the two different IBM Spectrum Scale
clusters.
The migration process of a file system managed by IBM Spectrum Archive used to be time
consuming because all the files on the source file system needed to be recalled from tapes
and copied to the target file system. From version 1.3.2.0, it supports a migration procedure
using SOBAR feature of IBM Spectrum Scale. By using SOBAR, the stub files on the source
file system are re-created on the target file system as stub files. The file system migration can
be completed without recalling/copying file data from tapes.
The procedure is referred to as System Migration using SOBAR feature and requires prior
review by IBM. For more information, see this IBM Documentation web page.
Chapter 6. Managing daily operations of IBM Spectrum Archive Enterprise Edition 231
232 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 233
7.14, “Overlapping IBM Spectrum Scale policy rules” on page 248
7.15, “Storage pool assignment” on page 250
7.16, “Tape cartridge removal” on page 250
7.17, “Reusing LTFS formatted tape cartridges” on page 251
7.18, “Reusing non-LTFS tape cartridges” on page 253
7.19, “Moving tape cartridges between pools” on page 254
7.20, “Offline tape cartridges” on page 254
7.21, “Scheduling reconciliation and reclamation” on page 255
7.22, “License Expiration Handling” on page 255
7.23, “Disaster recovery” on page 256
7.24, “IBM Spectrum Archive EE problem determination” on page 261
7.25, “Collecting IBM Spectrum Archive EE logs for support” on page 262
7.26, “Backing up files within file systems that are managed by IBM Spectrum Archive EE”
on page 264
7.27, “IBM TS4500 Automated Media Verification with IBM Spectrum Archive EE” on
page 266
7.28, “How to disable commands on IBM Spectrum Archive EE” on page 271
7.29, “LTO 9 Media Optimization” on page 272
Important: All of the command examples in this chapter use the command without the full
file path name because we added the IBM Spectrum Archive EE directory
(/opt/ibm/ltfsee/bin) to the PATH variable of the operating system.
234 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.1 Preventing migration of the .SPACEMAN and metadata
directories
This section describes an IBM Spectrum Scale policy rule that you should have in place to
help ensure the correct operation of your IBM Spectrum Archive EE system.
You can prevent migration of the .SPACEMAN directory and the IBM Spectrum Archive EE
metadata directory of a IBM General Parallel File System (GPFS) file system by excluding
these directories by having an IBM Spectrum Scale policy rule in place. Example 7-1 shows
how an exclude statement can look in an IBM Spectrum Scale migration policy file where the
metadata directory starts with the text “/ibm/glues/.ltfsee”.
Example 7-1 IBM Spectrum Scale sample directory exclude statement in the migration policy file
define(
user_exclude_list,
(
PATH_NAME LIKE '/ibm/glues/.ltfsee/%'
OR PATH_NAME LIKE '%/.SpaceMan/%'
OR PATH_NAME LIKE '%/.snapshots/%'
)
)
For more information and detailed examples, see 6.11.2, “Threshold-based migration” on
page 168 and 6.11.3, “Manual migration” on page 173.
For example, if there are four drives and two copies initially, set the mount limit to 2 for the
primary tape cartridge pool and 2 for the copy tape cartridge pool. These settings maximize
the migration performance because both the primary and copy jobs are run in parallel by
using two tape drives each for each tape cartridge pool. This action also avoids unnecessary
mounts/unmounts of tape cartridges.
To show the current mount limit setting for a tape cartridge pool, run the following command:
eeadm pool show <poolname> [-l <libraryname>] [OPTIONS]
To set the mount limit setting for a tape cartridge pool, run the following command:
eeadm pool set <poolname> [-l <libraryname>] -a <attribute> -v <value>
236 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.3 Changing the SSH daemon settings
The default values for MaxSessions and MaxStartups are too low and must increase to allow
for successful operations with IBM Spectrum Archive EE. MaxSessions specifies the
maximum number of open sessions that is permitted per network connection. The default is
10.
Note: If SSH is slow, this might indicate that several things might be wrong. Often disabling
GSSAPI Authentication and reversing DNS lookup will resolve this problem and speed up
SSH. Thus, set the following lines in the sshd_config file:
GSSAPIAuthentication no
UseDNS no
Note: When using -N option, specify only the node class defined for IBM Spectrum Archive
EE nodes. This does not apply for cases where -I defer or -I prepare option is used.
-g GlobalWorkDirectory
Specifies a global work directory in which one or more nodes can store temporary files
during mmapplypolicy command processing. For more information about specifying more
than one node to process the command, see the description of the -N option. For more
information about temporary files, see the description of the -s option.
The global directory can be in the file system that mmapplypolicy is processing or in
another file system. The file system must be a shared file system, and it must be mounted
and available for reading and writing by every node that will participate in the
mmapplypolicy command processing.
If the -g option is not specified, then the global work directory is the directory that is
specified by the sharedTmpDir attribute of the mmchconfig command. If the sharedTmpDir
attribute is not set to a value, then the global work directory depends on the file system
format version of the target file system:
238 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
– If the target file system is at file system format version 5.0.1 or later (file system format
number 19.01 or later), the global work directory is the .mmSharedTmpDir directory at
the root level of the target file system.
– If the target file system is at a file system format version that is earlier than 5.0.1 then
the command does not use a global work directory.
If the global work directory that is specified by -g option or by the sharedTmpDir attribute
begins with a forward slash (/) then it is treated as an absolute path. Otherwise it is treated
as a path that is relative to the mount point of the file system or the location of the directory
to be processed.
If both the -g option and the -s option are specified, then temporary files can be stored in
both the specified directories. In general, the local work directory contains temporary files
that are written and read by a single node. The global work directory contains temporary
files that are written and read by more than one node.
If both the -g option and the -N option are specified, then mmapplypolicy uses
high-performance, fault-tolerant protocols during execution.
Note: It is always a preferred practice to specify the temp directory to something other than
/tmp in case the temporary files get large. This can be the case in large file systems, and
the use of the IBM Spectrum Scale file system is suggested.
7.5 Preferred inode size for IBM Spectrum Scale file systems
When you create the GPFS file systems, an option is available that is called -i InodeSize for
the mmcrfs command. The option specifies the byte size of inodes. By default, the inode size
is 4 KB and it consists of a fixed 128 byte header, plus data, such as disk addresses pointing
to data, or indirect blocks, or extended attributes.
The supported inode sizes are 512, 1024, and 4096 bytes. Regardless of the file sizes, the
preferred inode size is 4096 for all IBM Spectrum Scale file systems for IBM Spectrum
Archive EE. This size should include user data file systems and the IBM Spectrum Archive EE
metadata file system.
7.6 Determining the file states for all files within the GPFS file
system
Typically, to determine the state of a file and to which tape cartridges the file is migrated, you
run the eeadm file state command. However, it is not practical to run this command for
every file on the GPFS file system.
In Example 7-3, the file is in the migrated state and is only on the tape cartridge JD0321JD.
Example 7-4 Sample set of list rules to display the file states
define(
user_exclude_list,
(
PATH_NAME LIKE '/ibm/glues/.ltfsee/%'
OR PATH_NAME LIKE '%/.SpaceMan/%'
OR PATH_NAME LIKE '%/lost+found/%'
OR NAME = 'dsmerror.log'
)
)
define(
is_premigrated,
(MISC_ATTRIBUTES LIKE '%M%' AND MISC_ATTRIBUTES NOT LIKE '%V%')
)
define(
is_migrated,
(MISC_ATTRIBUTES LIKE '%V%')
)
define(
is_resident,
(NOT MISC_ATTRIBUTES LIKE '%M%')
)
define(
is_symlink,
(MISC_ATTRIBUTES LIKE '%L%')
)
define(
is_dir,
(MISC_ATTRIBUTES LIKE '%D%')
)
240 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
SHOW('premigrated ' || xattr('dmapi.IBMTPS'))
WHERE is_premigrated
The policy runs a script that is named file_states.sh, which is shown in Example 7-5. If the
policy is run daily, this script can be modified to keep several versions to be used for history
purposes.
To run the IBM Spectrum Scale policy, run the mmapplypolicy command with the -P option
and the file states policy. This action produces a file that is called /root/file_states.txt, as
shown in Example 7-6.
In the /root/file_states.txt file, the files states and file system objects can be easily
identified for all IBM Spectrum Scale files, including the tape cartridges where the files or file
system objects are.
When vm.min_free_kbytes is set to its default value, some configurations might encounter
memory exhaustion symptoms when free memory should in fact be available. Setting
vm.min_free_kbytes to a higher value of 5-6% of the total amount of physical memory, up to a
max of 2 GB, helps to avoid such a situation.
Inodes are allocated when they are used. When a file is deleted, the inode is reused, but
inodes are never deallocated. When setting the maximum number of inodes in a file system,
there is an option to preallocate inodes. However, in most cases there is no need to
preallocate inodes because by default inodes are allocated in sets as needed.
If you do decide to preallocate inodes, be careful not to preallocate more inodes than will be
used. Otherwise, the allocated inodes unnecessarily consume metadata space that cannot
be reclaimed.
242 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Consider the following points when managing inodes:
For file systems that are supporting parallel file creates, as the total number of free inodes
drops below 5% of the total number of inodes, there is the potential for slowdown in the file
system access. Take this situation into consideration when creating or changing your file
system.
Excessively increasing the value for the maximum number of inodes might cause the
allocation of too much disk space for control structures.
To view the current number of used inodes, number of free inodes, and maximum number of
inodes, run the following command:
mmdf Device
To set the maximum inode limit for the file system, run the following command:
mmchfs Device --inode-limit MaxNumInodes[:NumInodesToPreallocate]
For example, a file system block size should be 2 MB due to a disk subsystem of eight data
disks plus one parity disk with a stripe size of 256 KB.
Refer to IBM Spectrum Scale Documentation for more details on cache related parameters
such as maxFilesToCache and maxStatCache.
Note: In the following policies, some optional attributes are added to provide efficient
(pre)migration such as the SIZE attribute. This attribute specifies how many files to pass in
to the EXEC script at a time. The preferred setting, which is listed in the following examples,
is to set it to 20 GiB.
Example 7-7 shows a simple migration policy that chooses files greater than 5 MB to be
candidate migration files and stubs them to tape. This is a good base policy that you can
modify to your specific needs. For example, if you need to have files on three storage pools,
modify the OPTS parameter to include a third <pool>@<library>.
244 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.10.2 Creating active archive system policies
An active archive system requires two policies to maintain the system. The first is a
premigration policy that selects all files over 5 MB to premigrate to tape, allowing users to still
quickly obtain their files from disk. To see how to place this premigration policy into a cron job
to run every 6 hours, see “Using a cron job” on page 176.
Example 7-8 shows a simple premigration policy for files greater than 5 MB.
The second policy is a fail-safe policy that needs to be set so when a low disk space event is
triggered, the fail-safe policy can be called. Adding the WEIGHT attribute to the policy enables
the user to choose whether they want to start stubbing large files first or least recently used
files. When the fail-safe policy starts running, it frees up the disk space to a set percentage.
The following commands are used for setting a fail-safe policy and calling mmadcallback:
mmchpolicy gpfs failsafe_policy.txt
mmaddcallback MIGRATION --command /usr/lpp/mmfs/bin/mmapplypolicy --event
lowDiskSpace --parms “%fsName -B 20000 -m <2x the number of drives>
--single-instance”
After setting the policy with the mmchpolicy command, run mmaddcallback with the fail-safe
policy. This policy runs periodically to check whether the disk space has reached the
threshold where stubbing is required to free up space.
246 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
SIZE(20971520)
7.11 Capturing a core file on Red Hat Enterprise Linux with the
Automatic Bug Reporting Tool
The Automatic Bug Reporting Tool (ABRT) consists of the abrtd daemon and a number of
system services and utilities to process, analyze, and report detected problems. The daemon
runs silently in the background most of the time, and springs into action when an application
crashes or a kernel fault is detected. The daemon then collects the relevant problem data,
such as a core file if there is one, the crashing application’s command-line parameters, and
other data of forensic utility.
For abrtd to work with IBM Spectrum Archive EE, two configuration directives must be
modified in the /etc/abrt/abrt-action-save-package-data.conf file:
OpenGPGCheck = yes/no
Setting the OpenGPGCheck directive to yes, which is the default setting, tells ABRT to
analyze and handle only crashes in applications that are provided by packages that are
signed by the GPG keys, which are listed in the /etc/abrt/gpg_keys file. Setting
OpenGPGCheck to no tells ABRT to detect crashes in all programs.
ProcessUnpackaged = yes/no
This directive tells ABRT whether to process crashes in executable files that do not belong
to any package. The default setting is no.
Here is an example of how rsyslog ommail can be used with IBM Spectrum Archive EE by
modifying the /etc/rsyslog.conf file:
If users want to send an email on all IBM Spectrum Archive EE registered error messages,
the regular expression is “GLES[A-Z][0-9]*E”, as shown in Example 7-11.
Example 7-11 Email for all IBM Spectrum Archive EE registered error messages
$ModLoad ommail
$ActionMailSMTPServer us.ibm.com
$ActionMailFrom ltfsee@ltfsee_host1.tuc.stglabs.ibm.com
$ActionMailTo ltfsee_user1@us.ibm.com
$template mailSubject,"LTFS EE Alert on %hostname%"
$template mailBody,"%msg%"
$ActionMailSubject mailSubject
:msg, regex, "GLES[A-Z][0-9]*E" :ommail:;mailBody
After a file is migrated to a tape cartridge pool and is in the migrated state, it cannot be
migrated to other tape cartridge pools (unless it is recalled back from physical tape to file
system space).
248 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Do not use overlapping IBM Spectrum Scale policy rules within different IBM Spectrum Scale
policy files that can select the same files for migration to different tape cartridge pools. If a file
was migrated, a later migration fails. The migration result for any file that already is in the
migrated state is fail.
In Example 7-12, an attempt is made to migrate four files to tape cartridge pool pool2. Before
the migration attempt, Tape ID JCB610JC is already in tape cartridge pool pool1, and Tape ID
JD0321JD in pool2 has one migrated and one pre-migrated file. The state of the files on these
tape cartridges before the migration attempt is shown by the eeadm file state command in
Example 7-12.
Example 7-12 Display the state of files by using the eeadm file state command
[root@saitama1 prod]# eeadm file state *.bin
Name: /ibm/gpfs/prod/fileA.ppt
State: migrated
ID: 11151648183451819981-3451383879228984073-1435527450-974349-0
Replicas: 1
Tape 1: JCB610JC@pool1@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/fileB.ppt
State: migrated
ID: 11151648183451819981-3451383879228984073-2015134857-974348-0
Replicas: 1
Tape 1: JCB610JC@pool1@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/fileC.ppt
State: migrated
ID: 11151648183451819981-3451383879228984073-599546382-974350-0
Replicas: 1
Tape 1: JD0321JD@pool2@lib_saitama (tape state=appendable)
Name: /ibm/gpfs/prod/fileD.ppt
State: premigrated
ID: 11151648183451819981-3451383879228984073-2104982795-3068894-0
Replicas: 1
Tape 1: JD0321JD@pool2@lib_saitama (tape state=appendable)
The mig scan list file that is used in this example contains these entries, as shown in
Example 7-13.
The attempt to migrate the files produces the results that are shown in Example 7-14.
The files on Tape ID JCB610JC (fileA.ppt and fileB.pdf) already are in tape cartridge pool
pool1. Therefore, the attempt to migrate them to tape cartridge pool pool2 produces a
migration result failed.
For the files on Tape ID JD0321JD, the attempt to migrate fileC.ppt file produces a
already_migrated or a migration result fail for some cases with multiple library environments
because the file is already migrated. Only the attempt to migrate the pre-migrated fileD.ppt
file succeeds. Therefore, one operation succeeded and three other operations result with
wrong_pool or already_migrated.
If you put different logical parts of an IBM Spectrum Scale namespace (such as the project
directory) into different LTFS tape cartridge pools, you can Normal Export tape cartridges that
contain only the files from that specific part of the IBM Spectrum Scale namespace (such as
project abc). Otherwise, you must first recall all the files from the namespace of interest
(such as the project directory of all projects), migrate the recalled files to an empty tape
cartridge pool, and then Normal Export that tape cartridge pool.
The concept of different tape cartridge pools for different logical parts of an IBM Spectrum
Scale namespace can be further isolated by using IBM Spectrum Archive node groups. A
node group consists of one or more nodes that are connected to the same tape library. When
tape cartridge pools are created, they can be assigned to a specific node group. For migration
purposes, it allows certain tape cartridge pools to be used with only drives within the owning
node group.
For more information, see 6.8.2, “Moving tape cartridges” on page 156, and “The eeadm
<resource type> --help command” on page 316.
250 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.16.1 Reclaiming tape cartridges before you remove or export them
To avoid failed recall operations, it is recommended that the tape cartridges are reclaimed
before removing or exporting a cartridge.
When a cartridge that can be appended is planned for removal, use one of the following
methods to perform reclaim before the removal:
Run the eeadm tape reclaim command before you remove it from the LTFS file system (by
running the eeadm tape unassign command)
Export it from the LTFS library (by running the eeadm tape export command. This
command will internally run the reclaim during its task.)
If tape cartridges are in a need_replace or a require_replace state, use the eeadm tape
replace command. This command will internally run the reclaim command during its
procedure.
The eeadm tape unassign --safe-remove can be used for cases where the replace command
fails. Note that the --safe-remove option will recall all the active files on the tape back to an
IBM Spectrum Scale file system that has adequate free space, and those files need to be
manually migrated again to a good tape.
7.16.2 Exporting tape cartridges before physically removing them from the
library
A preferred practice is always to export a tape cartridge before it is physically removed from
the library. If a removed tape cartridge is modified and then reinserted in the library,
unpredictable behavior can occur.
Because these tape cartridges still might contain data from the previous usage, IBM
Spectrum Archive EE recognizes the old content because LTFS is a self-describing format.
Before such tape cartridges can be reused within your IBM Spectrum Archive EE
environment, the data must be moved off the cartridge or deleted from the file system then
the cartridges must be reformatted before they are added to an IBM Spectrum Archive EE
tape cartridge pool. This task can be done by running the
eeadm tape reclaim or eeadm tape unassign -E commands. Note that the tapes removed
with the -E option will need a -f option when re-assigning with the eeadm tape assign
command.
Example 7-15 Display content of a used LTFS tape cartridge (non-IBM Spectrum Archive EE)
[root@ltfs97 ~]# ls -la /ltfs/153AGWL5
total 41452613
drwxrwxrwx 2 root root 0 Jul 12 2012 .
drwxrwxrwx 12 root root 0 Jan 1 1970 ..
-rwxrwxrwx 1 root root 18601 Jul 12 2012 api_test.log
-rwxrwxrwx 1 root root 50963 Jul 11 2012 config.log
-rwxrwxrwx 1 root root 1048576 Jul 12 2012 dummy.000
-rwxrwxrwx 1 root root 21474836480 Jul 12 2012 perf_fcheck.000
-rwxrwxrwx 1 root root 20971520000 Jul 12 2012 perf_migrec
lrwxrwxrwx 1 root root 25 Jul 12 2012 symfile ->
/Users/piste/mnt/testfile
You can also discover if it was an IBM Spectrum Archive EE tape cartridge before or just a
standard LTFS tape cartridge that is used by IBM Spectrum Archive LE or IBM Spectrum
Archive SDE release. Review the hidden directory .LTFSEE_DATA, as shown in Example 7-16.
This example indicates that this cartridge was previously used as an IBM Spectrum Archive
EE tape cartridge.
Example 7-16 Display content of a used LTFS tape cartridge (IBM Spectrum Archive EE)
[root@ltfs97 ltfs]# ls -lsa /ltfs/JD0321JD
total 0
0 drwxrwxrwx 4 root root 0 Jan 9 14:33 .
0 drwxrwxrwx 7 root root 0 Dec 31 1969 ..
0 drwxrwxrwx 3 root root 0 Jan 9 14:33 ibm
0 drwxrwxrwx 2 root root 0 Jan 10 16:01 .LTFSEE_DATA
The procedure for reuse and reformatting of a previously used LTFS tape cartridge depends
on whether it was used before as an IBM Spectrum Archive LE or IBM Spectrum Archive SDE
tape cartridge or as an IBM Spectrum Archive EE tape cartridge.
Before you start with the reformat procedures and examples, it is important that you confirm
the following starting point. You can see the tape cartridges that you want to reuse by running
the eeadm tape list command in the status unassign, as shown in Example 7-17.
252 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Reformatting and reusing an LTFS SDE/LE tape cartridge
In this case, you run the eeadm tape assign command to add this tape cartridge to an IBM
Spectrum Archive EE tape cartridge pool and format it at the same time, as shown in the
following example and in Example 7-18:
eeadm tape assign <list_of_tapes> -p <pool> [OPTIONS]
If the format fails then there is data most likely written to the tape and a force format is
required by appending the -f option to the eeadm tape assign command.
Although these tape cartridges still might contain data from the previous usage, they can be
used within IBM Spectrum Archive EE the same way as new, unused tape cartridges. For
more information about how to add new tape cartridge media to an IBM Spectrum Archive EE
tape cartridge pool, see 6.8.1, “Adding tape cartridges” on page 154.
7.19.1 Avoiding changing assignments for tape cartridges that contain files
If a tape cartridge contains any files, a preferred practice is to not move the tape cartridge
from one tape cartridge pool to another tape cartridge pool. If you remove the tape cartridge
from one tape cartridge pool and then add it to another tape cartridge pool, the tape cartridge
includes files that are targeted for multiple pools. This is not internally allowed in IBM
Spectrum Archive EE.
Before you export files you want from that tape cartridge, you must recall any files that are not
supposed to be exported in such a scenario.
For more information, see 6.9, “Tape storage pool management” on page 162.
For more information, see 6.9, “Tape storage pool management” on page 162 and 6.17,
“Reclamation” on page 200.
254 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.21 Scheduling reconciliation and reclamation
This section provides information about scheduling regular reconciliation and reclamation
activities.
The reconciliation process resolves any inconsistencies that develop between files in the IBM
Spectrum Scale and their equivalents in LTFS. The reclamation function frees up tape
cartridge space that is occupied by non-referenced files and non-referenced content that is
present on the tape cartridge. In other words, this is inactive content of data, but still
occupying space on the physical tape.
For more information, see 6.15, “Recalling files to their resident state” on page 196 and 6.17,
“Reclamation” on page 200.
When the license is expired and detected by the scheduler of the main IBM Spectrum Archive
EE management components (MMM), it shuts down. This feature is necessary to have a
proper clean-up if some jobs are still running or unscheduled. By doing so, a user is aware
that IBM Spectrum Archive EE does not function because of the license expiration.
To give a user the possibility to access files that were previously migrated to tape, it is
possible for IBM Spectrum Archive EE to restart, but it operates with limited functions. All
functions that write to tape cartridges are not available. During the start of IBM Spectrum
Archive EE (through MMM), it is detected that some nodes have the status of
NODE_STATUS_LICENSE_EXPIRED.
In 1992, the SHARE user group in the United States, along with IBM, defined a set of DR tier
levels. This action was done to address the need to describe and quantify various different
methodologies for successful mission-critical computer systems DR implementations. So,
within the IT Business Continuance industry, the tier concept continues to be used, and is
useful for describing today’s DR capabilities.
The tiers’ definitions are designed so that emerging DR technologies can also be applied, as
listed in Table 7-1.
3 Electronic vaulting
256 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
In the context of the IBM Spectrum Archive EE product, this section focuses only on the tier 1
strategy because this is the only supported solution that you can achieve with a product that
handles physical tape media (off-site vaulting).
For more information about the other DR tiers and general strategies, see Disaster Recovery
Strategies with Tivoli Storage Management, SG24-6844.
This environment can also establish a backup platform, although it does not have a site at
which to restore its data, nor the necessary hardware on which to restore the data, such as
compatible tape devices.
Because vaulting and retrieval of data is typically handled by couriers, this tier is described as
the PTAM. PTAM is a method that is used by many sites because this is a relatively
inexpensive option. However, it can be difficult to manage because it is difficult to know
exactly where the data is at any point.
There is probably only selectively saved data. Certain requirements were determined and
documented in a contingency plan and there is optional backup hardware and a backup
facility that is available. Recovery depends on when hardware can be supplied, or possibly
when a building for the new infrastructure can be located and prepared.
Although some customers are on this tier and seemingly can recover if there is a disaster, one
factor that is sometimes overlooked is the recovery time objective (RTO). For example,
although it is possible to recover data eventually, it might take several days or weeks. An
outage of business data for this long can affect business operations for several months or
even years (if not permanently).
Important: With IBM Spectrum Archive EE, the recovery time can be improved because
after the import of the vaulting tape cartridges into a recovered production environment,
the user data is immediately accessible without the need to copy back content from the
tape cartridges into a disk or file system.
The fundamental concept is based on the IBM Spectrum Archive EE function to create
replicas and redundant copies of your file system data to tape media during migration (see
6.11.4, “Replicas and redundant copies” on page 177). IBM Spectrum Archive EE enables
the creation of a replica plus two more redundant replicas (copies) of each IBM Spectrum
Scale file during the migration process.
The first replica is the primary copy, and other replicas are called redundant copies.
Redundant copies must be created in tape cartridge pools that are different from the tape
cartridge pool of the primary copy and different from the tape cartridge pools of other
redundant copies.
Up to two redundant copies can be created, which means that a specific file from the GPFS
file system can be stored on three different physical tape cartridges in three different IBM
Spectrum Archive EE tape cartridge pools.
The tape cartridge where the primary copy is stored and the tapes that contain the redundant
copies are referenced in the IBM Spectrum Scale inode with an IBM Spectrum Archive EE
DMAPI attribute. The primary copy is always listed first.
Redundant copies are written to their corresponding tape cartridges in the IBM Spectrum
Archive EE format. These tape cartridges can be reconciled, exported, reclaimed, or
imported by using the same commands and procedures that are used for standard migration
without replica creation.
Redundant copies must be created in tape cartridge pools that are different from the pool of
the primary copy and different from the pools of other redundant copies. Therefore, create a
DR pool named DRPool that exclusively contains the media you plan to Offline Export for
offline vaulting. You must also plan for the following issues:
Which file system data is migrated (as another replica) to the DR pool?
How often do you plan to export and remove of physical tapes for offline vaulting?
How do you handle media lifecycle management with the tape cartridges for offline
vaulting?
What are the DR steps and procedure?
If the primary copy of the IBM Spectrum Archive EE server and IBM Spectrum Scale do not
exist due to a disaster, the redundant copy that is created and stored in an external site
(offline vaulting) will be used for the disaster recovery.
Section 7.23.3, “IBM Spectrum Archive EE tier 1 DR procedure” on page 259 describes the
steps that are used to perform these actions:
Recover (import) the offline vaulting tape cartridges to a newly installed IBM Spectrum
Archive EE environment
Re-create the GPFS file system information
Regain access to your IBM Spectrum Archive EE data
Important: The migration of a pre-migrated file does not create new replicas.
258 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 7-19 shows you a sample migration policy to migrate all files to three pools. To have
this policy run periodically, see “Using a cron job” on page 176.
After you create redundant copies of your file system data on different IBM Spectrum Archive
EE tape cartridge pools for offline vaulting, you can Normal Export the tape cartridges by
running the IBM Spectrum Archive EE export command. For more information, see 6.19.2,
“Exporting tape cartridges” on page 204.
Important: The IBM Spectrum Archive EE export command does not eject the tape
cartridge to the physical I/O station of the attached tape library. To eject the DR tape
cartridges from the library to take them out for offline vaulting, you can run the eeadm tape
move command with the option -L ieslot. For more information, see 6.8, “Tape library
management” on page 154 and 10.1, “Command-line reference” on page 316.
The eeadm tape import command reinstantiates the stub files in IBM Spectrum Scale for
migrated files. The state of those files changes to the migrated state. Also, the eeadm tape
import command re-creates the file system objects in IBM Spectrum Scale for saved file
system objects.
Note: When a symbolic link is saved to tape and then restored by the eeadm tape import
command, the target of the symbolic link is kept. However, this process might cause the
link to break. Therefore, after a symbolic link is restored, it might need to be moved
manually to its original location on IBM Spectrum Scale.
1 Note that only exported tapes are valid for this command.
260 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Import processing for unexported tapes that are not reconciled
The eeadm tape import command might encounter tapes that are not reconciled when the
command is applied to tapes that are not exported from IBM Spectrum Archive EE. In this
case, the following situations can occur with the processing to restore files and file system
objects, and should be handled as described for each case:
The tapes might have multiple generations of a file or a file system object. If so, the
eeadm tape import command restores an object from the latest one that is on the tapes
that are specified by the command.
The tapes might not reflect the latest file information from IBM Spectrum Scale. If so, the
eeadm tape import command restores files or file system objects that were removed from
IBM Spectrum Scale.
For more information about the eeadm tape import command, see “Importing” on page 202.
The IBM Spectrum Archive EE log file can be found in the following directory:
/var/log/ltfsee.log
For more information about problem determination, see Chapter 9, “Troubleshooting IBM
Spectrum Archive Enterprise Edition” on page 293.
2. Open /etc/rsyslog.conf, and in the Global Directives section, add the following lines:
$imjournalRatelimitInterval 0
$imjournalRatelimitBurst 0
A Linux script is available with IBM Spectrum Archive EE that collects all of the needed files
and logs for your convenience to provide them to IBM Support. This task also compresses the
files into a single package.
To generate the compressed .tar file and provide it on request to IBM Support, run the
following command:
ltfsee_log_collection
262 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 7-21 shows the output of the ltfsee_log_collection command. During the log
collection run, you are asked what information you want to collect. If you are unsure, select Y
to select all the information. At the end of the output, you can find the file name and where the
log package was stored.
This program collects the following information from your IBM Spectrum Scale (GPFS)
cluster.
(1) Log files that are generated by IBM Spectrum Scale (GPFS), and IBM Spectrum Archive
EE
(2) Configuration information that is configured to use IBM Spectrum Scale (GPFS)
and IBM Spectrum Archive EE.
(3) System information including
OS distribution and kernel information,
hardware information (CPU and memory) and
process information (list of running processes).
(4) Task information files under the following subdirectory
<GPFS mount point>/.ltfsee/statesave
The following files are collected only if they were modified within the last 90 days.
- /var/log/messages*
- /var/log/ltfsee.log*
- /var/log/ltfsee_trc.log*
- /var/log/ltfsee_mon.log*
- /var/log/ltfs.log*
- /var/log/ltfsee_install.log
- /var/log/ltfsee_stat_driveperf.log*
- /var/log/httpd/error_log*
- /var/log/httpd/rest_log*
- /var/log/ltfsee_rest/rest_app.log*
- /var/log/logstash/*
You can collect all of the above files, including files modified within the last 90 days,
with an argument of 'all'.
#./ltfsee_log_collection all
If you want to collect the above files that were modified within the last 30 days.
#./ltfsee_log_collection 30
Input > y
Creating a temporary directory '/root/ltfsee_log_files'...
The collection of local log files is in progress.
The mmbackup command utilizes all the scalable, parallel processing capabilities of the
mmapplypolicy command to scan the file system, evaluate the metadata of all the objects in
the file system, and determine which files need to be sent to backup in IBM Spectrum Protect,
as well which deleted files should be expired from IBM Spectrum Protect. Both backup and
expiration take place when running mmbackup in the incremental backup mode.
The mmbackup command can inter-operate with regular IBM Spectrum Protect commands for
backup and expire operations. However if after using mmbackup, any IBM Spectrum Protect
incremental or selective backup or expire commands are used, mmbackup needs to be
informed of these activities. Use either the -q option or the --rebuild option in the next
mmbackup command invocation to enable mmbackup to rebuild its shadow databases.
These databases shadow the inventory of objects in IBM Spectrum Protect so that only new
changes will be backed up in the next incremental mmbackup. Failing to do so will needlessly
back up some files again. The shadow database can also become out of date if mmbackup fails
due to certain IBM Spectrum Protect server problems that prevent mmbackup from properly
updating its shadow database after a backup. In these cases it is also required to issue the
next mmbackup command with either the -q option or the --rebuild options.
264 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
7.26.1 Considerations
Consider the following points when you are using the IBM Spectrum Protect Backup/Archive
client in the IBM Spectrum Archive EE environment:
IBM Spectrum Protect requirements for backup
Update the dsm.sys and dsm.opt files to support both IBM Spectrum Protect and IBM
Spectrum Archive EE operations:
Note: In dsm.sys file, when multiple server stanza are defined for IBM Spectrum
Archive and IBM Spectrum Protect, the following lines need to be placed at the
beginning of the file. Otherwise, IBM Spectrum Archive will not be able to migrate or
recall files.
HSMBACKENDMODE TSMFREE
ERRORLOGNAME /opt/tivoli/tsm/client/hsm/bin/dsmerror.log
MAXRECALLDAEMONS 64
MINRECALLDAEMONS 64
HSMMIGZEROBLOCKFILES YES
errorlogretention 180 S
Ensure the files are backed up first with IBM Spectrum Protect followed by archiving the
files with IBM Spectrum Archive EE to avoid recall storms.
The mmbackup command is used to back up the files of a GPFS or IBM Spectrum Scale file
system to IBM Spectrum Protect servers by using the IBM Spectrum Protect Backup/Archive
Client of the IBM Spectrum Protect family. In addition, the mmbackup command can operate
with regular IBM Spectrum Protect backup commands for backup. After a file system is
backed up, you can restore files by using the interfaces that are provided by the IBM
Spectrum Protect family.
Starting with IBM Spectrum Archive EE v1.3.0.0 and IBM Spectrum Scale v5.0.2.2, a new
option has been added to the eeadm migrate command called --mmbackup. With the
--mmbackup option is supplied, IBM Spectrum Archive EE will first verify that current backup
versions of the files exist within IBM Spectrum Protect before it will archive them to IBM
Spectrum Archive EE. If those files are not backed up, those files will be filtered out and thus
not archived to tape. This ensures that there will be no recall storm due to the backup of files.
The --mmbackup option of the eeadm migrate command takes 1 option, which is the location of
the mmbackup shadow database. This location is normally the same device or directory
option of the mmbackup command.
Starting with the release of the IBM TS4500 Tape Library R2, a new, fully transparent function
is introduced within the TS4500 operations that is named policy-based automatic media
verification. This new function is hidden from any ISV software, similar to the automatic
cleaning.
No ISV certification is required. It can be enabled/disabled through the logical library with
more settings to define the verify period (for example, every 6 months) and the first
verification date.
One or more designated media verification drives (MVDs) must be assigned to a logical
library in order for the verification to take place. A preferred practice is to have two MVDs
assigned at a time to ensure that no false positives occur because of a faulty tape drive.
Figure 7-2 shows an example of such a setup.
Figure 7-2 TS4500 with one logical library showing two MVDs configured
Note: MVDs defined within a logical library are not accessible from the host or application
by using the drives of this particular logical library for production.
266 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Verify that results are simple pass/fail, but verify that failures are retried on a second physical
drive, if available, before being reported. A failure is reported through all normal notification
options (email, syslog, and SNMP). MVDs are not reported as mount points (SCSI DTEs) to
the ISV application, so MVDs do not need to be connected to the SAN.
During this process, whenever access from the application or host is required to the physical
media under media verification, the tape library stops the current verification process. It then
dismounts the needed tape from the MVD, and mounts it to a regular tape drive within the
same logical library for access by the host application to satisfy the requested mount. At a
later point, the media verification process continues.
The library GUI Cartridges page adds columns for last verification date/time, verification
result, and next verification date/time (if automatic media verification is enabled). If a cartridge
being verified is requested for ejection or mounting by the ISV software (which thinks the
cartridge is in a storage slot), the verify task is automatically canceled, a checkpoint occurs,
and the task resumes later (if/when the cartridge is available).
The ISV eject or mount occurs with a delay comparable to a mount to a drive being cleaned
(well within the preferred practice SCSI Move Medium timeout values). The GUI also supports
a manual stop of the verify task.
The last verification date/time is written in the cartridge memory (CM) and read upon first
mount after being newly inserted into a TS4500, providing persistence and portability (similar
to a cleaning cartridge usage count).
All verify mounts are recorded in the mount history CSV file, allowing for more granular health
analysis (for example, outlier recovered error counts) by using Tape System Reporter (TSR)
or Rocket Server graph.
The whole media verification process is transparent to IBM Spectrum Archive EE as the host.
No definitions and configurations need to be done within IBM Spectrum Archive EE. All setup
activities are done only through the TS4500 management interface.
Figure 7-3 Assign a tape drive to a logical library through the TS4500 web interface (step 1)
2. Another window opens where you must select the specific logical library to which the
unassigned drive is supposed to be added, as shown in Figure 7-4.
Figure 7-4 Assign a tape drive to a logical library through the TS4500 web interface (step 2)
268 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
3. If the drive to be used as an MVD is configured within the logical library, change its role, as
shown in Figure 7-5 and Figure 7-6 on page 269.
Figure 7-5 Reserve a tape drive as the media verification drive through the TS4500 web interface
4. You must right-click the assigned drive within the logical library. A menu opens and you
select Use for Media Verification from the list of the provided options. A confirmation
dialog box opens. Click Yes to proceed.
5. After making that configuration change to the drive, you see a new icon in front of it to
show you the new role (Figure 7-6).
Figure 7-6 Display a tape drive as the media verification drive through the TS4500 web interface
Note: The MVD flag for a tape drive is a global setting, which means that after it is
assigned, the drive keeps its role as an MVD even it is unassigned and then assigned to a
new logical library. Unassigning does not disable this role.
To unassign a drive from being an MVD, follow the same procedure again, and select (after
the right-click) Use for Media Access. This action changes the drive role back to normal
operation for the attached host application to this logical library.
By using this dialog box, you can enable/disable an automatic media verification schedule.
Then, you can configure how often the media should be verified and the first verification date.
Finally, you can select the MVDs, which are selected by the library to perform the scheduled
media verification test.
If you go to the Cartridges by Logical Library page and select Properties for the selected
logical library, a dialog box opens where you can see the current media verification
configuration for that logical library, as shown by Figure 7-8.
270 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
For more information and the usage of the TS4500 media verification functions, see IBM
TS4500 R8 Tape Library Guide, SG24-8235 and IBM TS4500 documentation at IBM
Documentation.
The control is done by disabling or enabling corresponding task types by using the eeadm
cluster set command. The commands, corresponding task types, and corresponding
attribute names that are used by the eeadm cluster set command are listed in Table 7-2.
The attributes can be set to “yes” or “no” by using the eeadm cluster set command. The
current setting can be verified by using the eeadm cluster show command.
When a task type is disabled, the next command immediately fails. The failed task can be
verified by using the eeadm task list -c command and the eeadm task show command.
For more information about LTO 9 media optimization, see this web page.
When using LTO 9 drives/cartridges in IBM Spectrum Archive managed systems, the media
optimization is performed as part of eeadm tape assign command. When deploying LTO-9
cartridges, the extra time the command may take due to the media optimization needs to be
accounted for.
272 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
One method for deploying multiple LTO 9 tapes into a large tape library managed by IBM
Spectrum Archive is as follows:
1. Use all of the available LTO 9 drives to perform the eeadm tape assign command to
prepare enough LTO 9 tapes for several days.
2. Use IBM Spectrum Archive drive role feature to dedicate one or a few drives to keep
performing the eeadm tape assign command while using other drives for migration and
recall. The drive role can be set by the eeadm drive set command.
– Set "g" to the drives that perform tape assign
– Set "mr" to the other drives for migration/recall
Also, the eeadm tape reclaim command formats the source tape at the last phase of the
command. If the source tape is LTO 9, the format will trigger the media optimization.
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 275
8.1 Overview of use cases
The typical use cases for IBM Spectrum Archive EE can be broken into three categories:
Archive, tiered storage, and data exchange, as shown in Figure 8-1.
Leverage simplicity and tape TCO Leverage import and export functions,
copy function, and standardized format
For more information about each use case, see Figure 8-2, Figure 8-3 on page 277, and
Figure 8-4 on page 278.
The solution is based on archive storage that is based on IBM Spectrum Scale, IBM
Spectrum Archive EE, and standard file system interfaces
276 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Some of the archive use case benefits are:
Simplicity with file system interface
Scalable with IBM Spectrum Scale and IBM Spectrum Archive EE
Low TCO with IBM tape
Some of the requirements for the tiered and scalable use case are:
Archive to file systems
Simple backup solution
Easy data access for restore
The solution is based on archive storage that is based on IBM Spectrum Scale, IBM
Spectrum Archive EE, and standard file system interfaces
Some of the requirements for the data exchange use case are:
Export entire directories to tape
Import files and directories with seamless data access
Leverage global name space
The solution is based on IBM Spectrum Scale, IBM Spectrum Archive EE, and standard file
system interfaces using export and import functions
278 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
8.2 Media and Entertainment
As visual effects get more sophisticated, the computational demands of media and
entertainment rose sharply. In response, Pixit Media became the United Kingdom’s leading
provider of post-production software-defined solutions that are based on IBM Spectrum
Storage technology, which provides reliable performance and cost-efficiency to support
blockbuster growth.
Their business challenge is to keep audiences spellbound with their output. Media and
Entertainment companies must share, store, and access huge files that traditional
infrastructure solutions are failing to deliver. To meet this transformation, Pixit Media put
consistent performance and chart-topping scalability in the limelight through software-defined
solutions that are based on IBM Spectrum Storage technology that helps customers create hit
after hit.
Figure 8-5 shows how a studio’s workflow and applications can talk to IBM Spectrum Scale,
which is backed by industry-standard servers and disk arrays that serve as the global single
namespace file system.
Pixit Media also began offering solutions that are based on IBM Spectrum Archive EE. They
commented that:
“Most of our clients have complex requirements when it comes to archiving a project. They
want to move it to a reliable media on a project-by-project basis but without having to
manually manage the library. IBM Spectrum Archive EE can offer these customers a
worry-free, centralized approach to managing this process, which can be set up in just two or
three days and scale out extremely quickly.”1
1 https://www.ibm.com/case-studies/pixit-media-systems-spectrum
If the post production videos are ever needed again, they can transparently be recalled back
to the high-speed storage. The user data files can be viewed as a normal file system to the
Media Asset Management (MAM) system while providing a seamless integration into
environments. The mission is to preserve these assets and provide rapid access. Figure 8-6
shows an IBM Spectrum Archive use case for media and entertainment.
Figure 8-6 IBM Spectrum Archive use case for media and entertainment
280 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
8.4 High-Performance Computing
Institutions and universities have long used High-Performance Computing (HPC) resources to
gather, store, and process their deep wealth of data which is often kept indefinitely. One key
step within this process is the ability to easily transfer, share, and discuss the data within their
own research teams and others. But this is often a major hurdle for the researchers.
Data transfers, with Globus, overcome this hurdle by providing a secure and unified interface
to their research data. Globus handles the complexities behind the details of large-scale data
transfers (such as tuning performance parameters), maintains security, monitors progress,
and validates correctness. Globus includes a “fire and forget” model where users submit the
data transfer task and are notified by Globus when the task is complete. Therefore,
researchers can concentrate only on performing their research.
Every data transfer task has a source endpoint and a destination endpoint. The endpoints are
described as the different locations where data can be moved to or from using the Globus
transfer, sync, and sharing service. For the IBM Spectrum Scale file system and IBM
Spectrum Archive EE tape usage, the following terminology is used:
Data is archived when you use the IBM Spectrum Scale file system as a destination
endpoint and it is cached until it is migrated to tape.
Data is restored or recalled when you use the IBM Spectrum Scale file system as a source
endpoint and might require a bulk recall of data from tape before the actual data transfers
can occur.
When recalls from tape through IBM Spectrum Archive EE are required, the ability to optimize
the bulk recalls is important because it can require a significant amount of time to complete.
Without optimizations, the tape recalls look randomized because there is no queue to be able
to group tape recalls which are on the same tape or within the same area of tape. Therefore,
a lot of time will be consumed by locating or rewinding times on the tape, and on the
unmounting or mounting of the target tape.
Therefore, starting with Globus Connect Server (GCS) version 5.4, Globus introduced a
feature called “posix staging” to allow the files to be prestaged to a disk cache before
performing the data transfer. With IBM Spectrum Scale and IBM Spectrum Archive EE, this
feature allows the optimization of bulk recalls from tape, prestaging them to the IBM Spectrum
Scale file system before being accessed by Globus. This process is done through Globus by
calling a specific “staging app” to generate the bulk recalls inside IBM Spectrum Archive EE.
A bulk recall can contain up to 64 staging requests per data transfer task.
For more information about the Globus posix staging feature, see this web page.
282 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
8.5 Healthcare
Amsterdam University Medical Center (UMC), location VUmc is enabling ground breaking
research with scalable, cost-effective storage for big data. With its storage requirements
skyrocketing that must be kept for decades, they needed to reevaluate its infrastructure to
continue conducting cutting-edge research in a secure and cost-effective way. Working with a
partner, the medical center helped researchers migrate from NAS drives to a centralized
storage platform that is based on IBM Spectrum Storage solutions.
With IBM Spectrum Scale and IBM Spectrum Archive EE solutions at the heart of its
centralized storage environment, Amsterdam UMC, location, VUmc supports its clinicians,
researchers, and administrators with the resources they need to work effectively. The promise
is to deliver a solution with which users can find the exact data they are looking for easily and
quickly, when and where they need it, and without disruptions.
When users or researchers have a new storage request, they are presented with a menu that
contains three storage options: gold, silver, and bronze. The gold option is on disk, while the
silver and bronze options are on tape (see Figure 8-8). Even when the data is on tape, it is
accessible from the online tape archive that provides the lower cost per TB and makes the
entire solution environmentally friendly and green.
Now the solution helps to achieve 99% faster data migrations, which enables IT to focus on
value-added development. The centralized data architecture ensures VUmc can fully support
its clinicians, researchers, and administrators with the resources they need to accelerate
discoveries and conduct innovative research.
The software engineers have also optimized the recall of massive genomics data for their
researchers, allowing for quick access to TBs of their migrated genomics data. Figure 8-9
shows a high-level archive for a genomics data archive.
Figure 8-9 IBM Spectrum Archive use case for genomics data archive
284 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
8.7 Archive of research and scientific data for extended periods
This research institute routinely manages large volumes of data generated internally and
collected by other institutes. To support its ongoing projects, the institute must archive and
store the data for many years. However, because most of the data is infrequently accessed,
the research institute was looking for a cost-efficient archiving solution that would allow
transparent user access.
In addition, the research institute needed a solution that would facilitate the fast import and
export of large data volumes. Figure 8-10 shows the high-level architecture to archive
research and scientific data for longs periods.
Figure 8-10 IBM Spectrum Archive use case for archiving research/scientific data for long periods of time
Figure 8-10 also shows the redundancy of the archive data from the backup solution. Both
archive and backup solutions are storing data on lower cost tape storage. The copies are in
two independent systems offering more options for stricter data security requirements.
The proposed solution will be a storage system residing in two locations, providing network
access to multiple organizational units within the university, each with their own respective
permission models. The primary objective of the Scientific Data Archive is to provide
cost-effective, resilient, long-term storage for research data and supporting research
computing infrastructure. The archive will be a storage facility operated the university’s
Storage Management Team, in cooperation with other units on campus. This facility will
deliver service to research organizations on campus. Figure 8-11 show the architecture for
the university’s archive.
Site 1 Site 2
Source Data, Database, Analytics Workstations Source Data, Database, Analytics Workstations
Figure 8-11 IBM Spectrum Archive use case for university Scientific Data Archive
286 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
8.9 Oil and gas
An oil and gas managed service provider collects offshore seismic data for analysis. The
solution uses the leading technology based on IBM Power Linux, IBM Spectrum Scale, and
IBM Spectrum Archive EE as a seismic data repository. The seismic data is collected from
vessels. The technology for both acquisition and processing of seismic data has evolved
dramatically over time. Today, a modern seismic vessel typically generates 4-5 TBs of new
raw data per day, which once processed will generate 10 to 100 times more data in different
formats.
For data that needs to be online but not accessed very frequently, tape is by far more
attractive than spinning disk. Hybrid storage solutions with automated and policy-driven
movement of data between different storage tiers including tape is required for such large
data repositories.
Figure 8-12 shows an IBM Spectrum Archive Oil and Gas archive use case.
Recall
segy
Figure 8-12 IBM Spectrum Archive use case for oil and gas
A video of every leading Managed Service Provider in the Nordics with thousands of servers
using tape as an integral part of their seismic data management is available at this website.
Therefore, as more and more companies leverage the advantage of object storage and
applications, the storage of the data becomes an important role. Data must be easily
accessed (regardless of protocol) and should not be tied to a specific tier. Access should be
in parallel, support tape throughput, and always on access.
IBM Spectrum Scale object storage combines the benefits of IBM Spectrum Scale with
OpenStack Swift, which is the most widely used open source object store. This object storage
system uses a distributed architecture with no central point of control, providing greater
scalability, redundancy, and the ability to access the objects via Swift API or the S3 API. But
IBM Spectrum Scale object storage requires the deployment of “Protocol Nodes” within the
IBM Spectrum Scale cluster.
Other than the IBM Spectrum Scale object storage, there is the option of using MinIO, which
provides the “S3 on Tape” concept. MinIO is a high-performance, open source, S3
compatible, enterprise hardened object storage. In NAS Gateway mode, MinIO provides the
S3-compatible environment that serves as the object storage endpoint. MinIO leverages the
IBM Spectrum Scale file system including multiple instances of MinIO NAS Gateways as a
distributed object storage.
The MinIO NAS Gateway is a simple translator for all S3 API calls, and writes the files on the
IBM Spectrum Scale file system as normal files. This MinIO NAS Gateway provides an
important feature called global 1-to-1 data sharing, which means that every object is a single
file on IBM Spectrum Scale (one object to one file). Any S3 object can be seen as a file
through IBM Spectrum Scale, including SMB/NFS, and any file created through IBM
Spectrum Scale, including SMB/NFS, can be seen as an object using the S3 API. You can
also create a separate NAS Gateway directory for only S3 data.
288 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
When you use MinIO with IBM Spectrum Scale and IBM Spectrum Archive EE, MinIO
handles all the S3 API calls while IBM Spectrum Scale provides the lifecycle management of
the objects to either remain on disk for hot or warm tier or on tape for the cold tier though the
powerful policy engine (see Figure 8-13).
Figure 8-13 High level MinIO with IBM Spectrum Archive architecture
To the S3 object user, all objects appear as they are in the object storage, available on disk,
and accessible using standard S3 GET or HEAD methods. Thus, it is simple to use it as any
standard S3 object storage, but it offers the advantage of tape economics for older objects
which have not been accessed for some time.
Active file management (AFM) uses a home-and-cache model in which a single home
provides the primary storage of data, and exported data is cached in a local GPFS file
system:
Home A home site is an NFS export of a remote cluster. This export can be a local file
system in the remote cluster, a GPFS file system, or a GPFS fileset in the remote
cluster. AFM is supported when a remote file system is mounted on the cache
cluster using GPFS protocols. This configuration requires that a multicluster setup
exists between the home and cache before AFM can use the home cluster’s file
system mount for AFM operations.
Cache A cache site is a remote cluster with a GPFS fileset that has a mount point to the
exported NFS file system of the home cluster’s file system. A cache site uses a
proprietary protocol over NFS. Each AFM-enabled fileset has a single home
cluster associated with it (represented by the host name of the home server).
In this setup, IBM Spectrum Archive EE is configured on the home cluster to archive all the
data generated from each cache cluster. The idea behind this solution is to have a single
home repository that has a large disk space, and multiple cache sites that cannot afford large
disk space.
Note: AFM supports multiple cache modes, and this solution can be used with single writer
or independent writer. However, with the release of IBM Spectrum Archive EE v1.2.3.0,
only the independent writer is currently supported.
When files are generated on the cache clusters, they are asynchronously replicated to the
home site. When these files are no longer being accessed on the cache clusters, the files can
be evicted, freeing up disk space at the cache clusters. They can then be migrated onto tape
at the home cluster. If evicted files need to be accessed again at the cache clusters, they can
simply be recovered by opening the file for access or by using AFM’s prefetch operation to
retrieve multiple files back to disk from the home site.
290 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Figure 8-14 shows a configuration of a single home cluster with multiple cache clusters to
form a centralized archive repository.
WAN
IBM Spectrum
Archive EE
Some examples of customers who can benefit from this solution are research groups that are
spread out geographically and rely on each group’s data, such as universities. Medical groups
and media companies can also benefit.
Asynchronous archive replication requires two remote clusters configured, one being the
home cluster and the other being a cache cluster with the independent writer mode. By using
the independent writer mode in this configuration, users can create files at either site and the
data/metadata is asynchronously replicated to the other site.
WAN
292 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
9
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 293
9.1 Overview
This section provides a simple health check procedure for IBM Spectrum Archive EE.
dsmmigfs query -N=all (verify that HSM status is active on every IBM Spectrum Archive node)
Yes
eeadminfo
ltfsee node list (check status of IBM Spectrum Archive EE nodes)
nodes
Status of all
nodes is
No
Correct the issue for the identified error module
Available?
Yes
All OK
If your issue remains after you perform these simple checks, follow the procedures that are
described in the remainder of this chapter to perform more detailed troubleshooting. If the
problem cannot be resolved, contact IBM Spectrum Archive Support.
After running eeadm cluster start and the returned status code is error, IBM Spectrum
Archive EE failed to start correctly and user actions are required to remedy the situation. To
view the type of error that has occurred during startup and which nodes are effected run the
eeadm node list command
294 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Failed startup caused by rpcbind
The rpcbind utility is required for MMM to function correctly, if rpcbind is not running IBM
Spectrum Archive EE will start up with errors and will be unusable until the issue has been
resolved. In most cases this is caused when rpcbind is not started up. If the server had been
recently powered down for maintenance and was started back up rpcbind might not start up
automatically.
Example 9-1 shows the output of a failed IBM Spectrum Archive EE startup due to rpcbind
not running.
Spectrum Archive EE service (MMM) for library lib_tora fails to start or is not
running on tora.tuc.stglabs.ibm.com Node ID:1
Problem Detected:
Node ID Error Modules
1 MMM; rpcbind;
To remedy this issue use the systemctl start rpcbind command to start the process up, and
either wait for the IBM Spectrum Archive EE Monitor daemon to start up MMM or issue a
eeadm cluster stop and eeadm cluster start to get the MMM started. After rpcbind is
started, it can be verified by running the systemctl status rpcbind to see if it is running.
Example 9-2 shows how to startup rpcbind and waiting for the Monitor daemon to restart
MMM.
Example 9-3 shows the output of ltfs -o device_list with no drives connected.
The second most common LE error is caused after connecting the drives to the IBM
Spectrum Archive EE node and forgetting to set at least one of the drives a control path drive
from the library GUI. IBM Spectrum Archive EE requires at least one control path drive so it
can communicate with the library.
Example 9-4 shows the output of a failed IBM Spectrum Archive EE start up that is caused by
LE.
296 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
.
Starting - waiting for getting ready to operate : lib_tora.
...
2019-01-21 09:24:33 GLESL657E: Fail to start the IBM Spectrum Archive EE service
(MMM) for library lib_tora.
Use the "eeadm node list" command to see the error modules.
The monitor daemon will start the recovery sequence.
[root@tora ~]# eeadm node list
Spectrum Archive EE service (MMM) for library lib_tora fails to start or is not
running on tora.tuc.stglabs.ibm.com Node ID:1
Problem Detected:
Node ID Error Modules
1 LE; MMM;
To remedy this failure, ensure that the node can see its drives and at least one drive is a
control path drive.
9.2 Hardware
This section provides information that can help you to identify and resolve problems with the
hardware that is used by IBM Spectrum Archive EE.
For more information about how to solve tape library errors, see the IBM TS4500 R8 Tape
Library Guide, SG24-8235.
If you right-click the event and select Display fix procedure, another window opens and
shows suggestions about how to fix the problem. If a drive display reports a specific drive
error code, see the tape drive maintenance manual for a solution, or call IBM H/W service.
For more information about analyzing the operating system error logs, see 9.5.1, “Linux” on
page 303.
If a problem is identified in the tape drive and the tape drive must be repaired, the drive must
first be removed from the IBM Spectrum Archive EE system. For more information, see
“Taking a tape drive offline” on page 299.
The size of each drive dump file is approximately 2 MB. By managing the drive dump files that
are generated, you can save disk space and enhance IBM Spectrum Archive performance.
298 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
It is not necessary to keep dump files after they are used for problem analysis. Likewise, the
files are not necessary if the problems are minor and can be ignored. A script program that is
provided with IBM Spectrum Archive EE periodically checks the number of drive dump files
and their date and time. If some of the dump files are older than two weeks or if the number of
dump files exceeds 1000 files, the script program erases them.
The script file is started by using Linux crontab features. A cron_ltfs_limit_dumps.sh file is
in the /etc/cron.daily directory. This script file is started daily by the Linux operating
system. The interval to run the script can be changed by moving the
cron_ltfs_limit_dumps.sh file to other cron folders, such as cron.weekly. For more
information about how to change the crontab setting, see the manual for your version of
Linux.
In the cron_ltfs_limit_dumps.sh file, the automatic drive dump erase policy is specified by
the option of the ltfs_limit_dump.sh script file, as shown in the following example:
/opt/ibm/ltfsle/bin/ltfs_limit_dumps.sh -t 14 -n 1000
You can modify the policy by editing the options in the cron_ltfs_limit_dumps.sh file. The
expiration date is set as a number of days by the -t option. In the example, a drive dump file
is erased when it is more than 14 days old. The number of files to keep is set by the -n option.
In our example, if the number of files exceeds 1,000, older files are erased so that the
1,000-file maximum is not exceeded. If either of the options are deleted, the dump files are
deleted by the remaining policy.
By editing these options in the cron_ltfs_limit_dumps.sh file, the number of days that files
are kept and the number of files that are stored can be modified.
Although not recommended, you can disable the automatic erasure of drive dump files by
removing the cron_ltfs_limit_dumps.sh file from the cron folder.
Important: If the diagnostic operation you intend to perform requires that a tape cartridge
be loaded into the drive, ensure that you have an empty non-pool tape cartridge available
in the logical library of IBM Spectrum Archive EE. If a tape cartridge is in the tape drive
when the drive is removed, the tape cartridge is automatically moved to the home slot.
3. Check the success of the removal. Run the eeadm drive list command and verify that
the output shows that the MMM attribute for the drive is in the stock state. Example 9-8
shows the status of the drives after it is removed from IBM Spectrum Archive EE.
4. Identify the primary device number of the drive for subsequent operations by running the
/opt/ibm/ltfsle/bin/ltfs -o device_list command. The command outputs a list of
available drives. Example 9-9 shows the output of this command.
300 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Device Name = /dev/sg0, Vender ID = IBM , Product ID = 0359260F ,
Serial Number = 000000014A00, Product Name =[0359260F].
5. If your diagnostic operations require a tape cartridge to be loaded into the drive, complete
the following steps. Otherwise, you are ready to perform diagnostic operations on the
drive, which has the drive address /dev/sgnumber, where number is the device number that
is obtained in step 4:
a. Move the tape cartridge to the drive from the I/O station or home slot. You can move
the tape cartridge by using ITDT (in which case the drive must have the control path),
or the TS4500 Management GUI.
b. Perform the diagnostic operations on the drive, which has the drive address
/dev/sgnumber, where number is the device number that is obtained in step 4.
c. When you are finished, return the tape cartridge to its original location.
6. Add the drive to the IBM Spectrum Archive EE inventory again by running the eeadm drive
assign drive_serial -n node_id command, where node_id is the same node that the
drive was assigned to originally in step 1 on page 299.
Example 9-10 shows the tape drive that is readded to IBM Spectrum Archive EE.
Running the eeadm drive list command again shows that the tape drive is no longer in a
“unassigned” state. Example 9-11 shows the output of this command.
In the rare case where the error on tape is permanent (several eeadm tape replace
commands have failed), it is suggested to try the eeadm tape unassigned --safe-remove
command instead. The --safe remove option recalls all the active files on the tape back to an
IBM Spectrum Scale file system that has adequate free space. The files then need to be
manually migrated again to a good tape again. Specify only one tape for use with the
--safe-remove option.
Complete the following steps to copy migrated jobs from a need_replace tape to a valid tape
within the same pool:
1. Identify a tape with a read failure by running eeadm tape list to locate the need_replace
tape.
2. Verify there is an appendable tape within the same pool which has enough available space
to hold the data from the need_replace tape cartridge.
3. Run the eeadm tape replace command to start the data transfer process onto an
appendable tape within the same pool and unassigning it from the tape cartridge pool at
the end.
4. Run eeadm tape move command on the need_replace tape to move the I/O station in order
to dispose of the bad tape.
302 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 9-13 shows system output of the steps to recover data from a read failure tape.
In the seldom case where the error on tape is permanent such that several eeadm tape
replace commands have failed, it is suggested to try the eeadm tape unassign
--safe-remove command instead. The --safe remove option recalls all the active files on the
tape back to an IBM Spectrum Scale file system that has adequate free space, and those files
need to be manually migrated again to a good tape again. Specify only one tape for the use
with --safe-remove option.
9.5 Software
IBM Spectrum Archive EE is composed of four major components, each with its own set of log
files. Therefore, problem analysis is slightly more involved than other products. This section
describes troubleshooting issues with each component in turn and the Linux operating
system and Simple Network Management Protocol (SNMP) alerts.
9.5.1 Linux
The log file /var/log/messages contains global LINUX system messages, including the
messages that are logged during system start and messages that are related to LTFS and
IBM Spectrum Archive EE functions. However, three specific log files are also created:
ltfs.log
ltfsee.log
ltfsee_trc.log
By default, after the ltfs.log, ltfsee.log, and ltfsee_trc.log files reach the threshold size,
they are rotated and four copies are kept. Example 9-15 shows the log file rotation settings.
These settings can be adjusted as needed within the /etc/logrotate.d/ibmsa-logrotate
control file.
304 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
These log files (ltfs.log, ltfsee.log, ltfsee_trc.log, and /var/log/messages) are
invaluable in troubleshooting LTFS messages. The ltfsee.log file contains only warning and
error messages. Therefore, it is easy to start looking here for the reasons of failure. For
example, a typical file migration might return the information message that is shown in
Example 9-16.
From the GLESL159E message, you know that the migration was unsuccessful, but you do
not know why it was unsuccessful. To understand why, you must examine the ltfsee.log file.
Example 9-17 shows the end of the ltfsee.log file immediately after the failed migrate
command is run.
In this case, the migration of the file was unsuccessful because it was previously
migrated/premigrated to a different tape pool.
With IBM Spectrum Archive EE, there are two logging facilities. One is in a human-readable
format that is monitored by users and the other is in machine-readable format that is used for
further problem analysis. The former facility is logged in to /var/log/ltfsee.log through the
“user” syslog facility and contains only warnings and errors. The latter facility is logged in to
/var/log/ltfsee_trc.log through the “local2” Linux facility.
The operating system automatically generates system log files after initial configuration of the
IBM Spectrum Archive EE. Provide the results of the ltfsee_log_collection command to
your IBM service representative.
The IBM Spectrum Scale log from the prior start of IBM Spectrum Scale can be found by
using the symbolic file name /var/adm/ras/mmfs.log.previous. All other files have a time
stamp and node name that is appended to the file name.
At IBM Spectrum Scale start, files that were not accessed during the last 10 days are deleted.
If you want to save old files, copy them elsewhere.
Example 9-18 shows normal operational messages that appear in the IBM Spectrum Scale
log file.
Example 9-18 Normal operational messages in an IBM Spectrum Scale log file
[root@ltfs97 ]# cat /var/adm/ras/mmfs.log.latest
Wed Apr 3 13:25:04 JST 2013: runmmfs starting
Removing old /var/adm/ras/mmfs.log.* files:
Unloading modules from /lib/modules/2.6.32-220.el6.x86_64/extra
Loading modules from /lib/modules/2.6.32-220.el6.x86_64/extra
Module Size Used by
mmfs26 1749012 0
mmfslinux 311300 1 mmfs26
tracedev 29552 2 mmfs26,mmfslinux
Wed Apr 3 13:25:06.026 2013: mmfsd initializing. {Version: 3.5.0.7 Built: Dec
12 2012 19:00:50} ...
Wed Apr 3 13:25:06.731 2013: Pagepool has size 3013632K bytes instead of the
requested 29360128K bytes.
Wed Apr 3 13:25:07.409 2013: Node 192.168.208.97 (htohru9) is now the Group
Leader.
Wed Apr 3 13:25:07.411 2013: This node (192.168.208.97 (htohru9)) is now Cluster
Manager for htohru9.ltd.sdl.
306 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Starting ADSM Space Management daemons
Wed Apr 3 13:25:17.907 2013: mmfsd ready
Wed Apr 3 13:25:18 JST 2013: mmcommon mmfsup invoked. Parameters: 192.168.208.97
192.168.208.97 all
Wed Apr 3 13:25:18 JST 2013: mounting /dev/gpfs
Wed Apr 3 13:25:18.179 2013: Command: mount gpfs
Wed Apr 3 13:25:18.353 2013: Node 192.168.208.97 (htohru9) appointed as manager
for gpfs.
Wed Apr 3 13:25:18.798 2013: Node 192.168.208.97 (htohru9) completed take over
for gpfs.
Wed Apr 3 13:25:19.023 2013: Command: err 0: mount gpfs
Wed Apr 3 13:25:19 JST 2013: finished mounting /dev/gpfs
Depending on the size and complexity of your system configuration, the amount of time to
start IBM Spectrum Scale varies. Taking your system configuration into consideration, if you
cannot access a file system that is mounted (automatically or by running a mount command)
after a reasonable amount of time, examine the log file for error messages.
The IBM Spectrum Scale log is a repository of error conditions that were detected on each
node, and operational events, such as file system mounts. The IBM Spectrum Scale log is the
first place to look when you are attempting to debug abnormal events. Because IBM
Spectrum Scale is a cluster file system, events that occur on one node might affect system
behavior on other nodes, and all IBM Spectrum Scale logs can have relevant data.
A common error that might appear when trying to mount GPFS is that it cannot read
superblock. Example 9-19 shows the output of the error when trying to mount GPFS.
The cause of this error and failure to mount GPFS is that the GPFS file system had dmapi
enabled, but the HSM process has not been started. To get around this error and successfully
mount GPFS, issue the systemctl start hsm command, and make sure it is running by
issuing systemctl status hsm. After HSM is running, wait for the recall processes to initiate.
This process can be viewed by issuing ps -afe | grep dsm. Example 9-20 shows output of
starting HSM, checking the status, and mounting GPFS.
If HSM is already running, double check if the dsmrecalld daemons are running by issuing ps
-afe | grep dsm. If no dsmrecalld daemons are running, start them by issuing dsmmigfs
start. After they have been started, GPFS can be successfully mounted.
The messages that are referenced in this section provide possible actions only for solvable
error codes. The error codes that are reported by LTFS program can be retrieved from the
terminal console or log files. For more information about retrieving error messages, see 9.5.1,
“Linux” on page 303.
When multiple errors are reported, LTFS attempts to find a message ID and an action for
each error code. If you cannot locate a message ID or an action for a reported error code,
LTFS encountered a critical problem. If you try an initial action again and continue to fail,
LTFS also encountered a critical problem. In these cases, contact your IBM service
representative for more support.
Message ID strings start with the keyword LTFS and are followed by a four- or five-digit value.
However, some message IDs include the uppercase letter I or D after LTFS, but before the
four- or five-digit value. When an IBM Spectrum Archive EE command is run and returns an
error, check the message ID to ensure that you do not mistake the letter I for the numeral 1.
A complete list of all LTFS messages can be found in the IBM Spectrum Archive EE section of
IBM Documentation.
At the end of the message ID, the following single capital letters indicate the importance of the
problem:
E: Error
W: Warning
I: Information
D: Debugging
308 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Example 9-21 shows a problem analysis procedure for LTFS.
The set of 10 characters represents the message ID, and the text that follows describes the
operational state of LTFS. The fourth message ID (LTFSI1079E) in this list indicates that an
error was generated because the last character is the letter E. The character immediately
following LTFS is the letter I. The complete message, including an explanation and
appropriate course of action for LTFSI1079E, is shown in the following example:
Example 9-22 on page 309.
Based on the description that is provided here, the tape cartridge in the library failed to
format. Upon further investigation the tape cartridge and drive is incompatible. The required
user action to solve the problem is to attach a compatible drive to the library and Spectrum
Archive EE node and rerun the operation.
The HSM log contains information about file migration and recall, threshold migration,
reconciliation, and starting and stopping the HSM daemon. You can analyze the HSM log to
determine the current state of the system. For example, the logs can indicate when a recall
has started but not finished within the last hour. The administrator can analyze a particular
recall and react accordingly.
In addition, an HSM log might be analyzed by an administrator to optimize HSM usage. For
example, if the HSM log indicates that 1,000 files are recalled at the same time, the
administrator might suggest that the files can be first compressed into one .tar file and then
migrated.
310 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
5. Send the tar.gz file to your IBM service representative.
Messages reference
For IBM Spectrum Archive EE, message ID strings start with the keyword GLES and are
followed by a single letter and then by a three-digit value. The single letter indicates which
component generated the message. For example, GLESL is used to indicate all messages
that are related to the IBM Spectrum Archive EE command. At the end of the message ID, the
following single uppercase letter indicates the importance of the problem:
E: Error
W: Warning
I: Information
D: Debugging
When you troubleshoot, check messages for errors only. For a list of available messages, see
IBM Documentation.
Failed reconciliations
Failed reconciliations usually are indicated by the GLESS003E error message with the
following description:
Reconciling tape %s failed due to a generic error.
File status
Table 9-1 lists the seven possible status codes for files in IBM Spectrum Archive EE. They can
be viewed for individual files by running the eeadm file state command.
resident The resident status indicates that the file is resident in the GPFS namespace
and is not saved, migrated, or premigrated to a tape.
migrated The migrated status indicates that the file was migrated. The file was copied from
GPFS file system to a tape, and exists only as a stub file in the GPFS
namespace.
premigrated The premigrated status indicates that the file was premigrated. The file was
copied to a tape (or tapes), but the file was not removed from the GPFS
namespace.
saved The saved status indicates that the file system object that has no data (a symbolic
link, an empty directory, or an empty regular file) was saved. The file system
object was copied from GPFS file system to a tape.
offline The offline status indicates that the file was saved or migrated to a tape
cartridge and thereafter the tape cartridge was exported offline.
missing The missing status indicates that a file has the migrated or premigrated status,
but it is not accessible from IBM Spectrum Archive EE because the tape cartridge
it is supposed to be on is not accessible. The file might be missing because of
tape corruption or if the tape cartridge was removed from the system without
exporting.
Files with the missing status are caused because of tape corruption or if the tape cartridge
was removed from the system without exporting. If the cause is a corrupted index run the
eeadm tape validate command, otherwise if the tape is missing from the tape library bring
the tape back into the library and run the eeadm library rescan command.
To recover a tape cartridge from a power failure, complete the following steps:
1. Create a mount point for the tape library. For more information, see the procedure
described in 6.2.2, “IBM Spectrum Archive Library Edition component” on page 137.
2. If you do not know which tape cartridges are in use, try to access all tape cartridges in the
library. If you do know which tape cartridges are in use, try to access the tape cartridge
that was in use when the power failure occurred.
3. If a tape cartridge is damaged, it is identified as inconsistent and the corresponding
subdirectories disappear from the file system. You can confirm which tape cartridges are
damaged or inconsistent by running the eeadm tape list command. The list of tape
cartridges that displays indicates the volume name, which is helpful in identifying the
inconsistent tape cartridge. For more information, see 6.18, “Checking and repairing
tapes” on page 201.
4. Recover the inconsistent tape cartridge by running the eeadm tape validate command.
For more information, see 6.18, “Checking and repairing tapes” on page 201.
Important: A drive in the library normally performs well despite a failure so that ongoing
access to an opened file on the loaded tape cartridge is not interrupted or damaged.
Important: One or more inconsistent tape cartridges might be found in the storage slots
and might need to be made consistent by following the procedure that is described in
“Unassigned state” on page 322.
312 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
If the user attempts to access the tape cartridge contents, the media folder is removed from
the file system. The status of any library tape cartridge can be determined by running the
eeadm tape list command. For more information, see 6.23, “Obtaining system resources,
and tasks information” on page 214.
Important: If the bar code is removed or about to peel off, the library cannot read it.
Replace the label or firmly attach the bar code to fix the problem.
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 315
10.1 Command-line reference
This section describes the IBM Spectrum Archive EE commands, IBM Spectrum Scale
commands, and IBM Spectrum Protect space management commands.
The eeadm command is initially used to get the high-level view of all available resource types
or subcommands that are available. Then, following the remaining help guide to find the
desired command to run.
For example, the eeadm cluster command manages the IBM Spectrum Archive EE cluster.
To display the available actions for the eeadm cluster command, enter the following
command:
eeadm cluster --help
316 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
In this case, the following actions for eeadm cluster are available:
eeadm cluster failover
eeadm cluster set
eeadm cluster show
eeadm cluster start
eeadm cluster stop
eeadm cluster restore
Therefore, the eeadm <resource type> --help command should be the next help guide
procedure that is used to find the available actions of each resource type.
Therefore, the eeadm <resource type> <action> --help command should be your final help
guide procedures used to run the wanted commands.
Table 10-1 Status and state codes for eeadm drive list
Drive Drive state Description and next action
status
ok mounted IBM Spectrum Archive EE can access this tape by using the file system interface,
but no task is assigned to this drive, and a tape is in this drive.
ok standby IBM Spectrum Archive EE cannot access a tape by using the file system interface,
but the tape is in this drive. No task is assigned to this drive.
ok mounting IBM Spectrum Archive EE is opening a file system interface to access a tape on this
drive because of an assigned task.
ok unmounting IBM Spectrum Archive EE is closing a file system interface access to the tape in this
drive because of opening a file system access to another tape, or moving a tape to
the homeslot.
info locked IBM Spectrum Archive EE is processing a tape in this drive that has a need_unlock
status. The need_unlock state is resolved automatically by the IBM Spectrum
Archive EE system.
Next Action: Check the nodes in IBM Spectrum Archive EE by using the eeadm node
list command, and make the nodes available by using the eeadm node up
command.
info unassigned This drive is not a resource of IBM Spectrum Archive EE.
Next Action: Use the eeadm drive assign command to include this drive as a
resource of IBM Spectrum Archive EE.
info disabled The drive is assigned to a node. But the drive is not open and IBM Spectrum
Archive EE cannot use this drive.
Next Action: Use the eeadm drive up command to enable this drive. The drive can
be used if the command is successful.
Next Action: Check the drive error messages on the library operator panel or web
GUI.
error not_installed The library is configured, but the drive does not exist as part of the library.
Next Action: Check the drive error messages on the library operator panel or web
GUI.
error missing The drive is defined as a resource of IBM Spectrum Archive EE. But it is not found
in the library.
Next Action: There is no degradation. Use the eeadm drive unassign command to
clean up this entry.
error check_drive_path IBM Spectrum Archive EE cannot find any alternative path at path fail over.
Next Action: Check drive connectivity on the assigned node, and restore
connectivity by using the eeadm drive up command.
318 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The eeadm file state command displays the data placement of files. Each file is in a unique
state, as listed in Table 10-2.
Table 10-2 Status codes for the eeadm file state command
Status code Description
resident The resident status indicates that the file is resident in the GPFS namespace and is
not saved, migrated, or premigrated to a tape.
migrated The migrated status indicates that the file was migrated. The file was copied from
GPFS file system to a tape, and exists only as a stub file in the GPFS namespace.
premigrated The premigrated status indicates that the file was premigrated. The file was copied
to a tape (or tapes), but the content of the file has not been truncated.
saved The saved status indicates that the file system object that has no data (a symbolic
link, an empty directory, or an empty regular file) was saved. The file system object
was copied from GPFS file system to a tape.
offline The offline status indicates that the file was saved or migrated to a tape cartridge
and thereafter the tape cartridge was exported offline.
missing The missing status indicates that a file has the migrated or premigrated status, but it
is not accessible from IBM Spectrum Archive EE because the tape cartridge it is
supposed to be on is not accessible. The file might be missing because of tape
corruption or if the tape cartridge was removed from the system without exporting.
Table 10-3 Status codes for the eeadm node list command
Status code Description
Available The Available status indicates that the IBM Spectrum Archive LE component on this
node is available for operation.
License Expired The License Expired status indicates that the IBM Spectrum Archive LE component
on this node has an expired license.
Unknown The Unknown status indicates that the IBM Spectrum Archive LE component on this
node is inoperable, or the node is down.
Disconnected The Disconnected status indicates that the EE and LE component on this node
cannot communicate. The admin channel connection might be disconnected.
Not configured The Not configured status indicates that the work directory of LE is not correct.
Reconfigure the node by stopping IBM Spectrum Archive EE and running the
ltfsee_config -m ADD_NODE command on the node. Then, restart IBM Spectrum
Archive EE.
Error The Error status indicates that a critical component of EE is not functioning or has
lost communication.
append_fenced Valid Pool setting caused this tape to be fenced from appending data.
check_hba N/A Temporary flag because of HBA-related error on mount, unmount, load,
or unload of the tape to or from the tape drive.
check_tape_library Unusable Temporary flag because of failed mount, unmount, load, or unload of the
tape to or from the tape drive.
data_full Valid The tape is almost full (only metadata changes are accepted).
disconnected Disconnect The Linear Tape File System (LTFS) LE instance that has this tape is not
running.
duplicated Duplicate Two or more tapes have the same barcode label in the logical library.
full Valid The tape is full and no other changes or updates can be made.
inaccessible Inaccessible The tape library reported that the tape is inaccessible.
label_mismatch N/A The barcode and label data on tape are mismatched.
need_unlock Critical This is an intermittent state while the IBM Spectrum Archive Enterprise
Edition system is processing a permanent write error. This state is
automatically resolved by the IBM Spectrum Archive Enterprise Edition
system.
non_supported Not supported The tape is an unsupported media cartridge and should be removed
from the logical library.
require_replace Critical or Error or The tape is in a read-only state because of a write error.
Write Fenced
require_validate Unknown The tape status is unknown and a tape validation is needed.
320 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Table 10-5 lists the next available commands for each tape status.
tape datamigrate
tape reconcile
tape unassign
tape validate
tape replace
tape reclaim
Status
tape assign
tape import
tape export
tape offline
tape online
premigrate
tape move
migrate
recall
save
appendable X X X X X X X X X X X X X
append_fenced X X X X X X X X X
check_hba X X
check_key_server X X
check_tape_library X X
data_full X X X X X X X X X
disconnected
duplicated
exported X X
full X X X X X X X
inaccessible
label_mismatch X
missing
need_replace X X X X X
non_supported X
offline X X
recall_only X X X X X X X
require_replace X X X X X
require_validate X X X
unassigned X X X
unformatted X
Require_validate state
A tape cartridge in the state is a temporary condition that can be caused when the tape within
the pool becomes unknown to the system, or the metadata on the gpfs file system becomes
inconsistent with the data cartridge.
Require_replace state
This status is caused by physical errors on writing data into the tape. The tape becomes a
read-only tape and must be replaced as soon as possible. For more information about
recovery procedures, see 9.3, “Recovering data from a write failure tape” on page 301 and
9.4, “Recovering data from a read failure tape” on page 302.
Need_replace state
This status is caused when a tape drive is encounters an error trying to read data off the tape.
This tape also must be replaced as soon as possible to mitigate future failures. For more
information about for recovery procedures, see “Recovering data from a read failure tape” on
page 302.
Unassigned state
This status is caused by a tape cartridge being removed from tape pool. The process of
adding it to LTFS (see 6.8.1, “Adding tape cartridges” on page 154) changes the status back
to appendable. Therefore, this message requires no other corrective action.
Unformatted status
This status usually is observed when a scratch tape is added to LTFS without formatting it. It
can be fixed by removing by eeadm tape unassign command and formatting it with the eeadm
tape assign command, as described in 6.8.3, “Formatting tape cartridges” on page 158.
If the tape cartridge was imported from another system, the IBM LTFS Format Verifier can be
useful for checking the tape format. For more information about performing diagnostic tests
with the IBM LTFS Format Verifier, see 10.3.2, “Using the IBM LTFS Format Verifier” on
page 328.
Inaccessible status
This status is most often the result of a stuck tape cartridge. Removing the stuck tape
cartridge and then moving it back to its homeslot (as described in 6.8.2, “Moving tape
cartridges” on page 156) should correct the Inaccessible status.
Non-supported status
Only LTO-9, 8, M8, 7, 6, 5, and 3592 JB, JC, JD, JE, JK, JL, JM, JV, JY, and JZ tape
cartridges are supported by IBM Spectrum Archive EE. This “tape state” or “status” indicates
that the tape cartridge is not one of these types and must be removed from the tape library.
322 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
10.1.5 IBM Spectrum Scale commands
Use these commands to manage GPFS file systems that you use with your IBM Spectrum
Archive EE system.
The node on which the command is run must have the GPFS mounted. The node must be
able to run remote shell commands on any other node in the IBM Spectrum Scale cluster
without the use of a password and without producing any extraneous messages.
For more information, see “Requirements for administering a GPFS file system” in GPFS:
Administration and Programming Reference, SA23-2221. GPFS documentation is available
at the following IBM Documentation web pages:
IBM Spectrum Scale
GPFS
For more information about the mmapplypolicy command and GPFS or IBM Spectrum Scale,
see your GPFS or IBM Spectrum Scale documentation.
The node on which the command is run must have the GPFS mounted. The node must be
able to run remote shell commands on any other node in the IBM Spectrum Scale cluster
without the use of a password and without producing any extraneous messages. For more
information, see “Requirements for administering a GPFS file system” in GPFS:
Administration and Programming Reference, SA23-2221, and the websites that are
referenced in “The mmapplypolicy command” on page 323.
The GPFS mmgetstate -a command displays the nodes where GPFS is active.
IBM Spectrum Archive EE provides a limited description of the commands in this subset. For
more information about these commands, see the “HSM client command reference” topic and
other related topics at this IBM Documentation web page.
Important: The IBM Spectrum Archive EE license does not entitle customers to use any
IBM Spectrum Protect components or products other than IBM Spectrum Protect for Space
Management from the IBM Spectrum Protect family to migrate data to LTFS.
The following IBM Spectrum Protect for Space Management commands are compatible with
the IBM Spectrum Archive EE environment:
dsmmigfs
dsmls
dsmrecall
Important: The HSM daemons are started with the same environment as the dsmwatchd
watch daemon. Therefore, the dsm.opt and dsm.sys options files in the IBM Spectrum
Protect for Space Management default installation path /usr/tivoli/tsm/client/ba/bin
are used.
Before you use this command, complete the following steps to verify your IBM Spectrum
Scale settings:
1. Verify that the IBM Spectrum Scale is active.
2. Verify that the node that you want to run the command from belongs to the IBM Spectrum
Archive EE cluster, and has the file system mounted.
Important: To display the nodes where IBM Spectrum Scale is active, run the IBM
Spectrum Scale mmgetstate -a command. For more information, see 6.2.1, “IBM
Spectrum Scale” on page 136.
dsmrecall
To selectively recall migrated files to the local file system, run the dsmrecall command.
Space management must be active.
324 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The duration of the recall process depends on the size of the files that are being recalled. It
takes time to mount and spool the LTFS tape cartridge, and the data transfer time for a large
file can be considerable.
This command includes the dsmrecall gpfs_path syntax, as shown in the following example:
dsmrecall /ibm/ltfs/filename1.txt
IBM Spectrum Archive EE uses the following data and metadata formats for migrating files
from GPFS to LTFS:
GPFS data
Figure 10-1 shows an example of original GPFS data.
Figure 10-2 Example layout of migrated data and symbolic links on LTFS (on tapes)
Migrated data
Migrated data is saved as files under the .LTFSEE_DATA directory on each tape. The
.LTFSEE_DATA directory is placed directly under the tape root directory on each tape. These
data files are stored under unique ID (UID) based file names. A UID-based file name
consists of the cluster ID (CLID), file system ID (FSID), inode generation number (IGEN),
and inode number (INO). In this example, all of the files have the same CLID and the same
FSID because all of the files belong to the same GPFS file system.
Symbolic links
The GPFS directory structure is rebuilt on each tape under the tape root directory for all
the migrated files that the tape contains. For each data file, a symbolic link is created
under the original GPFS file name and location, which points to the corresponding data file
on the tape. When a symbolic link is created, a relative path to the target is used so that if
the IBM Spectrum Archive mount point changes, the link stays correct. For the file1
example in Figure 10-2, the following symbolic link that corresponds to file1 is created on
tapeA:
/ltfs/tapeA/gpfs/file1 → ../.LTFSEE_DATA/CLID-FSID-IGEN1-INO1
GPFS path as metadata in IBM Spectrum Archive
For each data file in IBM Spectrum Archive, an extended attribute is set that contains the
original GPFS file path. In the file1 example in Figure 10-2, the two following LTFS files
have the extended attribute gpfs.path set to the value (file1 GPFS path) /gpfs/file1:
– /ltfs/tapeA/.LTFSEE_DATA/CLID-FSID-IGEN1-INO1
– /ltfs/tapeB/.LTFSEE_DATA/CLID-FSID-IGEN1-INO1
326 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
The saved GPFS path gives you the ability to re-create the original IBM Spectrum Scale
namespace by using the reconciliation and export processes followed by the import
process. In case of disaster, the approximate IBM Spectrum Scale namespace can be
recovered from tapes because without reconciliation, not all file deletions and renames in
IBM Spectrum Scale are reflected on the migration target tapes. The saved path is also
used for removing stale symbolic links when a file is recalled into resident state.
To download the most recent version of the LTFS LFV, complete the following steps:
1. Open the following URL in your web browser:
http://www.ibm.com/support/fixcentral
2. Click Product Group → System Storage.
3. Click Product Family → Tape systems.
4. Click Product Type → Tape drivers and software.
5. Click Product → Linear Tape File System (LTFS) Format Verifier.
6. Select Installed version.
7. Select your operating system from the Platform menu.
8. Click Continue.
9. Narrow the search of available downloads according to your criteria (this step can be
skipped).
10.Click Continue to view the list of available downloads.
11.Select the version that you want to download.
12.To download the new version, follow the instructions on the Fix Central download page.
Latest levels: IBM maintains the latest levels of the LTFS LFV and information about using
the tool and related documentation on the IBM Fix Central website.
Important: LTFS LFV is not shipped with IBM Spectrum Archive EE, but is available as a
separate download.
328 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
To verify that media are compatible with LTFS, run lfv from the command line. Enter one of
the following commands: For Linux systems where the IBM Tape Device Driver is installed,
<target device> should be /dev/IBMtapeX, where X is the index of the tape device to use, as
shown in the following example:
./lfv -f /dev/IBMtape1
For Linux systems where no IBM Tape Device driver is installed, <target device> should
be /dev/sgX, where X is the index for the tape device to use, as shown in the following
example:
./lfv -f /dev/sg0
Important: The index for the target tape device in the previous examples is shown as 0. If
you are unsure which index value to use, run the ./lfv -s command to scan for all
attached tape devices.
330 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
Related publications
The publications that are listed in this section are considered suitable for a more detailed
description of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide more information about the topics in this
document. Some publications that are referenced in this list might be available in softcopy
only:
Active Archive Implementation Guide with IBM Spectrum Scale Object and IBM Spectrum
Archive, REDP-5237
IBM Tape Library Guide for Open Systems, SG24-5946
IBM TS4500 R8 Tape Library Guide, SG24-8235
You can search for, view, download, or order these documents and other Redbooks,
Redpapers, Web Docs, draft, and other materials, at this website:
http://www.redbooks.ibm.com/
Other publications
IBM Tape Device Drivers Installation and User’s Guide, GC27-2130, also is relevant as
further information source.
Online resources
The following websites are also relevant as further information sources:
Tip: To learn more about IBM Spectrum Archive EE and try it in a virtual tape library
environment, see IBM Spectrum Archive Enterprise Edition Fundamentals & Lab Access
(log in required).
© Copyright IBM Corp. 2015, 2020, 2022. All rights reserved. 331
IBM Spectrum Archive Library Edition at IBM Documentation:
https://www.ibm.com/docs/en/spectrum-archive-le
IBM Spectrum Archive Enterprise Edition Support:
https://www.ibm.com/support/home/product/5449353/IBM_Spectrum_Archive_Enterpris
e_Edition_(EE)
IBM Spectrum Scale at IBM Documentation:
https://www.ibm.com/docs/en/spectrum-scale
IBM Systems Storage interactive product guide:
https://www.ibm.com/downloads/cas/AXYWNN5Z
Linear Tape-Open Program:
https://www.lto.org/lto-generation-compatibility/
SNIA Linear Tape File System Format Specification:
http://snia.org/sites/default/files/LTFS_Format_2.2.0_Technical_Position.pdf
332 IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
SG24-8333-09
IBM Spectrm Archive Enterprise Edition V1.3.2.2: Installation and Configuration Guide
ISBN 0738460427
(0.5” spine)
0.475”<->0.873”
250 <-> 459 pages
Back cover
SG24-8333-09
ISBN 0738460427
Printed in U.S.A.
®
ibm.com/redbooks