Isilon OneFS 7.1 - GUI
Isilon OneFS 7.1 - GUI
Isilon OneFS 7.1 - GUI
OneFS
Version 7.1
CONTENTS
Chapter 1
17
Chapter 2
19
Chapter 3
33
CONTENTS
Chapter 4
Access zones
69
Chapter 5
75
CONTENTS
OneFS privileges.............................................................................. 82
Command-line interface privileges................................................... 84
Authentication.............................................................................................. 88
Supported authentication providers.................................................88
Authentication provider features...................................................... 89
LDAP................................................................................................ 89
Active Directory................................................................................ 90
NIS...................................................................................................90
File provider..................................................................................... 90
Local provider.................................................................................. 91
Managing access permissions.......................................................................91
Configure access management settings........................................... 91
Modify ACL policy settings................................................................92
ACL policy settings options.............................................................. 92
Update cluster permissions..............................................................97
Managing roles............................................................................................. 99
View roles........................................................................................ 99
View privileges.................................................................................99
Create a custom role...................................................................... 100
Modify a role.................................................................................. 100
Delete a custom role...................................................................... 100
Managing authentication providers............................................................. 100
Configure an LDAP provider............................................................101
Managing LDAP providers.............................................................. 102
Configure an Active Directory provider............................................105
Managing Active Directory providers.............................................. 106
Managing NIS providers................................................................. 108
Configuring file providers............................................................... 110
Managing file providers..................................................................111
Create a local user......................................................................... 114
Create a local group....................................................................... 115
Managing local users and groups...................................................115
Chapter 6
Identity management
121
Chapter 7
Auditing
133
Auditing overview........................................................................................134
Protocol audit events.................................................................................. 134
Supported event types................................................................................ 134
Supported audit tools................................................................................. 135
Enable system configuration auditing..........................................................136
Enable protocol access auditing..................................................................136
Auditing settings......................................................................................... 137
Integrating with the EMC Common Event Enabler.........................................137
Install CEE for Windows..................................................................138
Configure CEE for Windows............................................................ 139
OneFS 7.1 Web Administration Guide
CONTENTS
Chapter 8
File sharing
141
Chapter 9
Snapshots
171
CONTENTS
Chapter 10
193
Deduplication overview...............................................................................194
Deduplication jobs......................................................................................194
Data replication and backup with deduplication..........................................195
Snapshots with deduplication.....................................................................195
Deduplication considerations......................................................................195
Shadow store considerations...................................................................... 196
SmartDedupe license functionality..............................................................196
Managing deduplication............................................................................. 197
Assess deduplication space savings.............................................. 197
Specify deduplication settings....................................................... 197
View deduplication space savings..................................................198
View a deduplication report........................................................... 198
OneFS 7.1 Web Administration Guide
CONTENTS
Chapter 11
201
CONTENTS
Chapter 12
237
FlexProtect overview....................................................................................238
File striping................................................................................................. 238
Requested data protection.......................................................................... 238
FlexProtect data recovery.............................................................................239
Smartfail........................................................................................ 239
Node failures................................................................................. 239
Requesting data protection......................................................................... 240
Requested protection settings.....................................................................240
Requested protection disk space usage...................................................... 241
Chapter 13
NDMP backup
243
CONTENTS
Chapter 14
265
Chapter 15
Protection domains
279
CONTENTS
Chapter 16
Data-at-rest-encryption
283
Chapter 17
SmartQuotas
291
Chapter 18
Storage Pools
313
11
CONTENTS
Spillover..................................................................................................... 316
Node pools................................................................................................. 317
Manual node pool management.....................................................317
SSD pools................................................................................................... 318
Tiers............................................................................................................319
File pools.................................................................................................... 319
File pool policies......................................................................................... 319
Managing node pools..................................................................................320
Add or move node pools in a tier.................................................... 320
Change the name or requested protection of a node pool...............320
Managing tiers............................................................................................ 321
Create a tier................................................................................... 321
Rename a tier.................................................................................321
Delete a tier................................................................................... 321
Creating file pool policies............................................................................322
Add a file pool policy......................................................................323
File pool file-matching options....................................................... 324
Valid wildcard characters............................................................... 325
Default file pool requested protection settings............................... 326
Default file pool I/O optimization settings...................................... 327
Managing file pool policies......................................................................... 327
Configure default file pool policy settings.......................................328
Configure default file pool protection settings................................ 328
Configure default I/O optimization settings....................................328
Modify a file pool policy................................................................. 329
Copy a file pool policy.................................................................... 329
Prioritize a file pool policy.............................................................. 329
Use a file pool template policy....................................................... 330
Delete a file pool policy..................................................................330
SmartPools settings....................................................................... 331
Monitoring storage pools............................................................................ 332
Monitor storage pools.................................................................... 333
View unhealthy subpools............................................................... 333
View file pool job results................................................................ 333
Chapter 19
System jobs
335
CONTENTS
Chapter 20
Networking
347
Chapter 21
Hadoop
375
Chapter 22
Antivirus
389
OneFS 7.1 Web Administration Guide
13
CONTENTS
Chapter 23
iSCSI
403
CONTENTS
Chapter 24
VMware integration
423
Chapter 25
429
15
CONTENTS
16
CHAPTER 1
Introduction to this guide
17
Live Chat
Create a Service Request
Telephone Support
18
CHAPTER 2
Isilon scale-out NAS
19
Node
S-Series
X-Series
Use Case
IOPS-intensive applications
High-concurrency and throughput-driven workflows
Node
Function
20
Isilon cluster
An Isilon cluster consists of three or more hardware nodes, up to 144. Each node runs the
Isilon OneFS operating system, the distributed file-system software that unites the nodes
into a cluster. A clusters storage capacity ranges from a minimum of 18 TB to a maximum
of 15.5 PB.
Cluster administration
OneFS centralizes cluster management through a web administration interface and a
command-line interface. Both interfaces provide methods to activate licenses, check the
status of nodes, configure the cluster, upgrade the system, generate alerts, view client
connections, track performance, and change various settings.
In addition, OneFS simplifies administration by automating maintenance with a job
engine. You can schedule jobs that scan for viruses, inspect disks for errors, reclaim disk
space, and check the integrity of the file system. The engine manages the jobs to
minimize impact on the cluster's performance.
With SNMP versions 1, 2c, and 3, you can remotely monitor hardware components, CPU
usage, switches, and network interfaces. EMC Isilon supplies management information
bases (MIBs) and traps for the OneFS operating system.
OneFS also includes a RESTful application programming interfaceknown as the Platform
APIto automate access, configuration, and monitoring. For example, you can retrieve
performance statistics, provision users, and tap the file system. The Platform API
integrates with OneFS role-based access control to increase security. See the Isilon
Platform API Reference.
Quorum
An Isilon cluster must have a quorum to work properly. A quorum prevents data conflicts
for example, conflicting versions of the same filein case two groups of nodes become
unsynchronized. If a cluster loses its quorum for read and write requests, you cannot
access the OneFS file system.
For a quorum, more than half the nodes must be available over the internal network. A
seven-node cluster, for example, requires a four-node quorum. A 10-node cluster requires
a six-node quorum. If a node is unreachable over the internal network, OneFS separates
the node from the cluster, an action referred to as splitting. After a cluster is split, cluster
operations continue as long as enough nodes remain connected to have a quorum.
In a split cluster, the nodes that remain in the cluster are referred to as the majority
group. Nodes that are split from the cluster are referred to as the minority group.
Internal and external networks
21
When split nodes can reconnect with the cluster and resynchronize with the other nodes,
the nodes rejoin the cluster's majority group, an action referred to as merging.
A OneFS cluster contains two quorum properties:
u
By connecting to a node with SSH and running the sysctl command-line tool as root,
you can view the status of both types of quorum. Here is an example for a cluster that has
a quorum for both read and write operations, as the command's output indicates with a
1, for true:
sysctl efs.gmp.has_quorum
efs.gmp.has_quorum: 1
sysctl efs.gmp.has_super_block_quorum
efs.gmp.has_super_block_quorum: 1
Storage pools
Storage pools segment nodes and files into logical divisions to simplify the management
and storage of data.
A storage pool comprises node pools and tiers. Node pools group equivalent nodes to
protect data and ensure reliability. Tiers combine node pools to optimize storage by
need, such as a frequently used high-speed tier or a rarely accessed archive.
The SmartPools module groups nodes and files into pools. If you do not activate a
SmartPools license, the module provisions node pools and creates one file pool. If you
activate the SmartPools license, you receive more features. You can, for example, create
multiple file pools and govern them with policies. The policies move files, directories, and
file pools among node pools or tiers. You can also define how OneFS handles write
operations when a node pool or tier is full. SmartPools reserves a virtual hot spare to
reprotect data if a drive fails regardless of whether the SmartPools license is activated.
IP address pools
Within a subnet, you can partition a cluster's external network interfaces into pools of IP
address ranges. The pools empower you to customize your storage network to serve
different groups of users. Although you must initially configure the default external IP
subnet in IPv4 format, you can configure additional subnets in IPv4 or IPv6.
You can associate IP address pools with a node, a group of nodes, or NIC ports. For
example, you can set up one subnet for storage nodes and another subnet for accelerator
nodes. Similarly, you can allocate ranges of IP addresses on a subnet to different teams,
such as engineering and sales. Such options help you create a storage topology that
matches the demands of your network.
In addition, network provisioning rules streamline the setup of external connections.
After you configure the rules with network settings, you can apply the settings to new
nodes.
As a standard feature, the OneFS SmartConnect module balances connections among
nodes by using a round-robin policy with static IP addresses and one IP address pool for
each subnet. Activating a SmartConnect Advanced license adds features, such as
defining IP address pools to support multiple DNS zones.
Storage pools
23
Data-access protocols
With the OneFS operating system, you can access data with multiple file-sharing and
transfer protocols. As a result, Microsoft Windows, UNIX, Linux, and Mac OS X clients can
share the same directories and files.
OneFS supports the following protocols.
-
Protocol Description
-
SMB
Server Message Block gives Windows users access to the cluster. OneFS works with
SMB 1, SMB 2, and SMB 2.1. With SMB 2.1, OneFS supports client opportunity locks
(oplocks) and large (1 MB) MTU sizes. The default file share is /ifs.
NFS
The Network File System enables UNIX, Linux, and Mac OS X systems to remotely
mount any subdirectory, including subdirectories created by Windows users. OneFS
works with versions 2 through 4 of the Network File System protocol (NFSv2, NFSv3,
NFSv4). The default export is /ifs.
FTP
File Transfer Protocol lets systems with an FTP client connect to the cluster to exchange
files.
iSCSI
The Internet Small Computer System Interface protocol provides access to block
storage. iSCSI integration requires you to activate a separate license.
HDFS
The Hadoop Distributed File System protocol makes it possible for a cluster to work
with Apache Hadoop, a framework for data-intensive distributed applications. HDFS
integration requires you to activate a separate license.
HTTP
Hyper Text Transfer protocol gives systems browser-based access to resources. OneFS
includes limited support for WebDAV.
A file provider for accounts in /etc/spwd.db and /etc/group files. With the file
provider, you can add an authoritative third-party source of user and group
information.
You can manage users with different identity management systems; OneFS maps the
accounts so that Windows and UNIX identities can coexist. A Windows user account
managed in Active Directory, for example, is mapped to a corresponding UNIX account in
NIS or LDAP.
To control access, an Isilon cluster works with both the access control lists (ACLs) of
Windows systems and the POSIX mode bits of UNIX systems. When OneFS must
24
transform a file's permissions from ACLs to mode bits or from mode bits to ACLs, OneFS
merges the permissions to maintain consistent security settings.
OneFS presents protocol-specific views of permissions so that NFS exports display mode
bits and SMB shares show ACLs. You can, however, manage not only mode bits but also
ACLs with standard UNIX tools, such as the chmod and chown commands. In addition,
ACL policies enable you to configure how OneFS manages permissions for networks that
mix Windows and UNIX systems.
Access zones
OneFS includes an access zones feature. Access zones allow users from different
authentication providers, such as two untrusted Active Directory domains, to access
different OneFS resources based on an incoming IP address. An access zone can
contain multiple authentication providers and SMB namespaces.
RBAC for administration
OneFS includes role-based access control (RBAC) for administration. In place of a
root or administrator account, RBAC lets you manage administrative access by role.
A role limits privileges to an area of administration. For example, you can create
separate administrator roles for security, auditing, storage, and backup.
It is recommended that you do not save data to the root /ifs file path but in directories
below /ifs. The design of your data storage structure should be planned carefully. A
well-designed directory optimizes cluster performance and cluster administration.
Data layout
OneFS evenly distributes data among a cluster's nodes with layout algorithms that
maximize storage efficiency and performance. The system continuously reallocates data
to conserve space.
OneFS breaks data down into smaller sections called blocks, and then the system places
the blocks in a stripe unit. By referencing either file data or erasure codes, a stripe unit
helps safeguard a file from a hardware failure. The size of a stripe unit depends on the
file size, the number of nodes, and the protection setting. After OneFS divides the data
into stripe units, OneFS allocates, or stripes, the stripe units across nodes in the cluster.
When a client connects to a node, the client's read and write operations take place on
multiple nodes. For example, when a client connects to a node and requests a file, the
node retrieves the data from multiple nodes and rebuilds the file. You can optimize how
OneFS lays out data to match your dominant access patternconcurrent, streaming, or
random.
Structure of the file system
25
Writing files
On a node, the input-output operations of the OneFS software stack split into two
functional layers: A top layer, or initiator, and a bottom layer, or participant. In read and
write operations, the initiator and the participant play different roles.
When a client writes a file to a node, the initiator on the node manages the layout of the
file on the cluster. First, the initiator divides the file into blocks of 8 KB each. Second, the
initiator places the blocks in one or more stripe units. At 128 KB, a stripe unit consists of
16 blocks. Third, the initiator spreads the stripe units across the cluster until they span a
width of the cluster, creating a stripe. The width of the stripe depends on the number of
nodes and the protection setting.
After dividing a file into stripe units, the initiator writes the data first to non-volatile
random-access memory (NVRAM) and then to disk. NVRAM retains the information when
the power is off.
During the write transaction, NVRAM guards against failed nodes with journaling. If a
node fails mid-transaction, the transaction restarts without the failed node. When the
node returns, it replays the journal from NVRAM to finish the transaction. The node also
runs the AutoBalance job to check the file's on-disk striping. Meanwhile, uncommitted
writes waiting in the cache are protected with mirroring. As a result, OneFS eliminates
multiple points of failure.
Reading files
In a read operation, a node acts as a manager to gather data from the other nodes and
present it to the requesting client.
Because an Isilon cluster's coherent cache spans all the nodes, OneFS can store different
data in each node's RAM. By using the internal InfiniBand network, a node can retrieve
file data from another node's cache faster than from its own local disk. If a read operation
requests data that is cached on any node, OneFS pulls the cached data to serve it
quickly.
In addition, for files with an access pattern of concurrent or streaming, OneFS pre-fetches
in-demand data into a managing node's local cache to further improve sequential-read
performance.
Metadata layout
OneFS protects metadata by spreading it across nodes and drives.
Metadatawhich includes information about where a file is stored, how it is protected,
and who can access itis stored in inodes and protected with locks in a B+ tree, a
standard structure for organizing data blocks in a file system to provide instant lookups.
OneFS replicates file metadata across the cluster so that there is no single point of
failure.
Working together as peers, all the nodes help manage metadata access and locking. If a
node detects an error in metadata, the node looks up the metadata in an alternate
location and then corrects the error.
26
Striping
In a process known as striping, OneFS segments files into units of data and then
distributes the units across nodes in a cluster. Striping protects your data and improves
cluster performance.
To distribute a file, OneFS reduces it to blocks of data, arranges the blocks into stripe
units, and then allocates the stripe units to nodes over the internal network.
At the same time, OneFS distributes erasure codes that protect the file. The erasure codes
encode the file's data in a distributed set of symbols, adding space-efficient redundancy.
With only a part of the symbol set, OneFS can recover the original file data.
Taken together, the data and its redundancy form a protection group for a region of file
data. OneFS places the protection groups on different drives on different nodescreating
data stripes.
Because OneFS stripes data across nodes that work together as peers, a user connecting
to any node can take advantage of the entire cluster's performance.
By default, OneFS optimizes striping for concurrent access. If your dominant access
pattern is streaming--that is, lower concurrency, higher single-stream workloads, such as
with video--you can change how OneFS lays out data to increase sequential-read
performance. To better handle streaming access, OneFS stripes data across more drives.
Streaming is most effective on clusters or subpools serving large files.
Antivirus
Description
-
OneFS can send files to servers running the Internet Content Adaptation
Protocol (ICAP) to scan for viruses and other threats.
27
Feature
-
Description
-
Clones
OneFS enables you to create clones that share blocks with other files to save
space.
OneFS can back up data to tape and other devices through the Network Data
Management Protocol. Although OneFS supports both NDMP 3-way and 2way backup, 2-way backup requires an Isilon Backup Accelerator node.
Protection
domains
The following software modules also help protect data, but they require you to activate a
separate license:
Licensed
Feature
-
SyncIQ
Description
-
SyncIQ replicates data on another Isilon cluster and automates failover and
failback operations between clusters. If a cluster becomes unusable, you can
fail over to another Isilon cluster.
SnapshotIQ
You can protect data with a snapshota logical copy of data stored on a
cluster.
SmartLock
The SmartLock tool prevents users from modifying and deleting files. You can
commit files to a write-once, read-many state: The file can never be modified
and cannot be deleted until after a set retention period. SmartLock can help
you comply with Securities and Exchange Commission Rule 17a-4.
28
Data mirroring
You can protect on-disk data with mirroring, which copies data to multiple locations.
OneFS supports two to eight mirrors. You can use mirroring instead of erasure codes, or
you can combine erasure codes with mirroring.
Mirroring, however, consumes more space than erasure codes. Mirroring data three
times, for example, duplicates the data three times, which requires more space than
erasure codes. As a result, mirroring suits transactions that require high performance,
such as with iSCSI LUNs.
You can also mix erasure codes with mirroring. During a write operation, OneFS divides
data into redundant protection groups. For files protected by erasure codes, a protection
group consists of data blocks and their erasure codes. For mirrored files, a protection
group contains all the mirrors of a set of blocks. OneFS can switch the type of protection
group as it writes a file to disk. By changing the protection group dynamically, OneFS can
continue writing data despite a node failure that prevents the cluster from applying
erasure codes. After the node is restored, OneFS automatically converts the mirrored
protection groups to erasure codes.
VMware integration
OneFS integrates with several VMware products, including vSphere, vCenter, and ESXi.
For example, OneFS works with the VMware vSphere API for Storage Awareness (VASA) so
that you can view information about an Isilon cluster in vSphere. OneFS also works with
the VMware vSphere API for Array Integration (VAAI) to support the following features for
block storage: hardware-assisted locking, full copy, and block zeroing. VAAI for NFS
requires an ESXi plug-in.
With the Isilon for vCenter plug-in, you can backup and restore virtual machines on an
Isilon cluster. With the Isilon Storage Replication Adapter, OneFS integrates with the
Data mirroring
29
VMware vCenter Site Recovery Manager to recover virtual machines that are replicated
between Isilon clusters.
Software modules
You can access advanced features by activating licenses for EMC Isilon software
modules.
SmartLock
SmartLock protects critical data from malicious, accidental, or premature alteration
or deletion to help you comply with SEC 17a-4 regulations. You can automatically
commit data to a tamper-proof state and then retain it with a compliance clock.
SyncIQ automated failover and failback
SyncIQ replicates data on another Isilon cluster and automates failover and failback
between clusters. If a cluster becomes unusable, you can fail over to another Isilon
cluster. Failback restores the original source data after the primary cluster becomes
available again.
File clones
OneFS provides provisioning of full read/write copies of files, LUNs, and other
clones. OneFS also provides virtual machine linked cloning through VMware API
integration.
SnapshotIQ
SnapshotIQ protects data with a snapshota logical copy of data stored on a
cluster. A snapshot can be restored to its top-level directory.
SmartPools
SmartPools enable you to create multiple file pools governed by file-pool policies.
The policies move files and directories among node pools or tiers. You can also
define how OneFS handles write operations when a node pool or tier is full.
SmartConnect
If you activate a SmartConnect Advanced license, you can balance policies to evenly
distribute CPU usage, client connections, or throughput. You can also define IP
address pools to support multiple DNS zones in a subnet. In addition, SmartConnect
supports IP failover, also known as NFS failover.
InsightIQ
The InsightIQ virtual appliance monitors and analyzes the performance of your Isilon
cluster to help you optimize storage resources and forecast capacity.
30
Software modules
31
CHAPTER 3
General cluster administration
33
User interfaces
Depending on your preference, location, or task, OneFS provides several interfaces for
managing the EMC Isilon cluster.
-
Interface
OneFS web
administration
interface
34
Description
The browser-based OneFS web
administration interface provides
secure access with OneFS-supported
browsers. You can use this interface to
view robust graphical monitoring
displays and to perform clustermanagement tasks.
You can run OneFS isi commands in
the command-line interface to
configure, monitor, and manage the
cluster. Access to the command-line
interface is through a secure shell
(SSH) connection to any node in the
cluster.
Comment
The OneFS web administration
interface uses port 8080 as its
default port.
OneFS RESTful
Access to the
Namespace API
Interface
Description
Comment
Note
35
Licensing
Advanced cluster features are available when you activate licenses for OneFS software
modules. Each optional OneFS software module requires you to activate a separate
license.
For more information about the following optional software modules, contact your EMC
Isilon sales representative.
u
HDFS
InsightIQ
SmartConnect Advanced
SmartDedupe
SmartLock
SmartPools
SmartQuotas
SnapshotIQ
SyncIQ
iSCSI
License status
The status of a OneFS module license indicates whether the functionality provided by the
module are available on the cluster.
Licenses exist in one of the following states:
-
Status
Inactive
Description
The license has not been activated on the cluster. You cannot access the features
provided by the corresponding module.
Evaluation The license has been temporarily activated on the cluster. You can access the
features provided by the corresponding module for a limited period of time. After the
license expires, the features will become unavailable, unless the license is
reactivated.
Activated
The license has been activated on the cluster. You can access the features provided
by the corresponding module.
Expired
The evaluation license has expired on the cluster. You can no longer access the
features provided by the corresponding module. The features will remain
unavailable, unless you reactivate the license.
The following table describes what functionality is available for each license depending
on the license's status:
License
-
36
HDFS
Inactive
-
Clients cannot
access the cluster
through HDFS.
Evaluation/
Activated
You can configure
HDFS settings and
clients can access
Expired
-
License
-
Inactive
-
Evaluation/
Activated
the cluster through
HDFS.
Expired
-
InsightIQ
SmartPools
SmartConnect
Advanced
Client connections
are balanced by
using a round robin
policy. IP address
allocation is static.
Each external
network subnet can
be assigned only
one IP address pool.
SmartDedupe
You cannot
deduplicate data
with SmartDedupe.
SmartLock
License status
37
License
-
38
Inactive
-
Evaluation/
Activated
Expired
-
SnapshotIQ
SmartQuotas
SyncIQ
iSCSI
Clients cannot
access the cluster
through iSCSI.
License configuration
You can configure or unconfigure some OneFS module licenses.
You can configure a license by performing specific operations through the corresponding
module. Not all actions that require you to activate a license will configure the license.
Also, not all licenses can be configured. Configuring a license does not add or remove
access to any features provided by a module.
You can unconfigure a license only through the isi license unconfigure
command. You may want to unconfigure a license for a OneFS software module if, for
example, you enabled an evaluation version of a module but later decided not to
purchase a permanent license. Unconfiguring a module license does not deactivate the
license. Unconfiguring a license does not add or remove access to any features provided
by a module.
The following table describes both the actions that cause each license to be configured
and the results of unconfiguring each license:
-
License
HDFS
InsightIQ
Cause of configuring
Cannot configure this license.
Result of unconfiguring
No system impact.
No system impact.
No system impact.
SmartPools
Create a file pool policy (other than the OneFS deletes all file pool policies
default file pool policy).
(except the default file pool policy).
SmartConnect
SmartDedupe
No system impact.
SmartLock
No system impact.
SnapshotIQ
SmartQuotas
Create a quota.
No system impact.
SyncIQ
No system impact.
iSCSI
No system impact.
Activate a license
To access a OneFS module, you must activate a license.
Before you begin
Before you can activate a license, you must obtain a valid license key, and you must have
root user privileges on your cluster. To obtain a license key, contact your EMC Isilon sales
representative.
Procedure
1. Click Help > About This Cluster.
2. In the Licensed Modules section, click Activate license.
3. In the License key field, type the license key for the module that you want to enable.
License configuration
39
4. Read the end user license agreement, click I have read and agree, and then click
Submit.
Unconfigure a license
You can unconfigure a licensed module through the command-line interface.
You must have root user privileges on your Isilon cluster to unconfigure a module license.
This procedure is available only through the command-line interface (CLI).
Note
If you do not know the module name, run the isi license command for a list of
OneFS modules and their status.
OnesFS returns a confirmation message similar to the following text: The
SmartConnect module has been unconfigured. The license is
unconfigured, and any processes enabled for the module are disabled.
Certificates
You can renew the Secure Sockets Layer (SSL) certificate for the Isilon web administration
interface or replace it with a third-party SSL certificate.
All Platform API communication, which includes communication through the web
administration interface, is over SSL. You can replace or renew the self-signed certificate
with a certificate that you generate. To replace or renew an SSL certificate, you must be
logged on as root.
40
Procedure
1. Establish an SSH connection to any node in the cluster.
2. At the command prompt, run the following command to create the appropriate
directory.
mkdir /ifs/local/
3. At the command prompt, run the following command to change to the directory.
cd /ifs/local/
Description
Third-party
(public or
private) CAissued
certificate
41
The steps to view security details vary by browser. For example, in some browsers, you
can click the padlock icon in the address bar to view the security details for the web
page. Follow the steps that are specific to your browser.
In addition, you should add the following attributes to be sent with your certificate
request:
u
Cluster name
Character encoding
Email settings
SNMP monitoring
SupportIQ settings
43
cluster and domain time become out of sync by more than 4 minutes, OneFS generates
an event notification.
Note
If the cluster and Active Directory become out of sync by more than 5 minutes,
authentication will not work.
To summarize:
u
If no NTP server is configured but the cluster is joined to an Active Directory domain,
the cluster synchronizes with Active Directory every 6 hours.
If an NTP server is configured, the cluster synchronizes the time with the NTP server.
44
3. Click Submit.
Description
-
Notes
-
Requires authorization
of every node added to
the cluster
If you use the secure join mode, you cannot use the serial
console wizard option [2] Join an existing cluster to join a
node to the cluster. You must add the node from cluster by
using the web administration interface or the isi devices
45
Mode
-
Description
-
Notes
-
-a add -d <unconfigured_node_serial_no>
command in the command-line interface.
To enable access time tracking, click Enabled, and then specify in the Precision
fields how often to update the last-accessed time by typing a numeric value and by
selecting a unit of measure, such as Seconds, Minutes, Hours, Days, Weeks,
Months, or Years.
For example, if you configure a Precision setting of 1 day, the cluster updates the
last-accessed time once each day, even if some files were accessed more often
than once during the day.
3. Click Submit.
Cluster statistics
You can view performance, historical, and in-depth usage statistics for your EMC Isilon
cluster, and control the output for each mode of statistics reporting.
The isi statistics and isi status command-line tools include options for
querying and filtering the display of EMC Isilon cluster performance and usage statistics.
Performance monitoring
You can view cluster throughput either graphically and numerically for average and
maximum usage.
Performance information is monitored through the web administration interface, or
through the command-line interface by using the isi statistics command options.
You can view details about the input and output traffic to and from the cluster's file
system. You can also monitor throughput distribution across the cluster.
Advanced performance monitoring and analytics are available through the InsightIQ
module, which requires you to activate a separate license. For more information about
optional software modules, contact your EMC Isilon Storage Division sales
representative.
Cluster monitoring
You can monitor the health, performance, and status of your EMC Isilon cluster.
Information is available for individual nodes, including node-specific network traffic,
internal and external network interfaces, and details about node pools, tiers, and overall
cluster health. You can monitor the following areas of your EMC Isilon cluster health and
performance:
Node status
Health and performance statistics for each node in the cluster, including hard disk
drive (HDD) and solid-state drive (SSD) usage.
Client connections
Number of clients connected per node.
New events
List of event notifications generated by system events, including the severity, unique
instance ID, start time, alert message, and scope of the event.
Cluster size
Current view: Used and available HDD and SSD space and space reserved for the
virtual hot spare (VHS). Historical view: Total used space and cluster size for a oneyear period.
Cluster statistics
47
Status: To view details about a node, click the ID number of the node.
Client connection summary: To view a list of current connections, click Dashboard >
Cluster Overview > Client Connections Status.
New events: To view more information about an event, click View details in the
Actions column.
Cluster size: To switch between current and historical views, click Historical or
Current near the Monitoring section heading. In historical view, click Used or
Cluster size to change the display.
Cluster throughput (file system): To switch between current and historical views,
click Historical or Current next to the Monitoring section heading. To view
throughput statistics for a specific period within the past two weeks, click
Dashboard > Cluster Overview > Throughput Distribution.
Note
CPU usage: To switch between current and historical views, click Historical or
Current near the Monitoring section heading.
Note
You can hide or show a plot by clicking System, User, or Total in the chart legend.
To view maximum usage, next to Show, select Maximum.
48
Status: To view networks settings for a node interface or subnet or pool, click the
link in the Status area.
Client connections: To view current clients connected to this node, review the list
in this area.
Chassis and drive status: To view the state of drives in this node, review this area.
To view details about a drive, click the name link of the drive; for example, Bay1.
Node size: To switch between current and historical views, click Historical or
Current next to the Monitoring area heading. In historical view, click Used or
Cluster size to change the display accordingly.
Node throughput (file system): To switch between current and historical views,
click Historical or Current next to the Monitoring area heading. To view throughput
statistics for a period within the past two weeks, click Dashboard > Cluster
Overview > Throughput Distribution.
Note
CPU usage: To switch between current and historical views, click Historical or
Current next to the Monitoring area heading.
Note
You can hide or show a plot by clicking System, User, or Total in the chart legend.
To view maximum usage, next to Show, select Maximum.
49
Email
You can designate recipients and specify SMTP, authorization, and security settings.
You can specify batch email settings and the email notification template.
SupportIQ
You can specify a protocol that you prefer to use for notifications: HTTPS, SMTP, or
both.
SNMP trap
You can send SNMP traps to one or more network monitoring stations or trap
receivers. Each event can generate one or more SNMP traps. You can download
management information base files (MIBs) from the cluster at /usr/local/
share/snmp/mibs/. The ISILON-TRAP-MIB.txt file describes the traps that
the cluster can generate, and the ISILON-MIB.txt file describes the associated
varbinds that accompany the traps.
Note
Notification batch
mode
Custom notification
template
50
Option
-
Batch all
Description
-
Batch by severity
Batch by category
No batching
No custom notification
template is set
Coalesced events
Related or repeated events are grouped, or coalesced, into one event by the OneFS
system.
There are two types of coalesced events.
Group events
Group events are different types of events that are all related to a single problem. For
example, a single connection problem might generate the following events:
-
Event
Description
Because the events are all related to a single problem, OneFS creates a group event
and adds the related errors to that event. Instead of seeing four events, you will see
a single group event alerting you to storage transport problems. You can still view all
the grouped events individually if you choose.
This message is representative of group coalesced event output.
# isi events show 24.924
ID:
24.924
Type:
199990001
Severity:
critical
Value:
0.0
Message:
Disk Errors detected (Bay 1)
Node:
21
Lifetime:
Sun Jun 17 23:29:29 2012 - Now
Quieted:
Not quieted
Specifiers: disk: 35
val: 0.0
devid: 24
drive_serial: 'XXXXXXXXXXXXX'
lba: 1953520064L
lnn: 21
drive_type: 'HDD'
device: 'da1'
bay: 1
unit: 805306368
Coalesced by: -Coalescer Type: Group
Coalesced events:
ID
STARTED
ENDED SEV LNN MESSAGE
24.911 06/17 23:29 -- I
21 Disk stall: Bay 1, Type HDD, LNUM
35. Disk ...
24.912 06/17 23:29 -- I
21 Sector error: da1 block 1953520064
24.913 06/17 23:29 -- I
21 Sector error: da1 block 2202232
24.914 06/17 23:29 -- I
21 Sector error: da1 block 2202120
24.915 06/17 23:29 -- I
21 Sector error: da1 block 2202104
24.916 06/17 23:29 -- I
21 Sector error: da1 block 2202616
24.917 06/17 23:29 -- I
21 Sector error: da1 block 2202168
24.918 06/17 23:29 -- I
21 Sector error: da1 block 2202106
24.919 06/17 23:29 -- I
21 Sector error: da1 block 2202105
24.920 06/17 23:29 -- I
21 Sector error: da1 block 1048670
24.921 06/17 23:29 -- I
21 Sector error: da1 block 223
24.922 06/17 23:29 -- C
21 Disk Repair Initiated: Bay 1, Type
HDD, LNUM...
51
You can view coalesced events and details through the web administration interface or
the command-line interface.
Responding to events
You can view event details and respond to cluster events.
You can view and manage new events, open events, and recently ended events. You can
also view coalesced events and additional, more-detailed information about a specific
event. You also can quiet or cancel events.
52
To add a new notification rule for this event, in the Create Rule area, select Create
a new notification rule for event, click Submit, and then specify the settings for the
rule.
To add the settings of this event to an existing event notification rule, in the Create
Rule area, select Add to an existing notification rule, select the existing event
notification rule from the list, and then click Submit.
Event label
COALESCED: FIRST EVENT
Description
An event was tagged as a possible first event in a series of events
that can be coalesced. The first event label is a only a placeholder
for a potential parent coalescer event.
COALESCED
CREATOR EV COALID UPDATED A group was created and the placeholder first event label was
updated to include actual group information.
DROPPED
An event did not include any new information and was not stored
in the master event database.
FORWARDED_TO_MASTER
DB: STORED
DB: PURGED
53
If a new event of the same event type is triggered, it is a separate new event and
must be quieted.
Unquiet
Returns a quieted event to an unacknowledged state in the list of new events and
removes the event from the list of quieted events.
Cancel
Permanently ends an occurrence of an event. The system cancels an event when
conditions are met that end its duration, which is bounded by a start time and an
end time, or when you cancel the event manually.
Most events are canceled automatically by the system when the event reaches the end of
its duration. The event remains in the system until you manually acknowledge or quiet
the event. You can acknowledge events through either the web administration interface
or the command-line interface.
Manage an event
You can change the status of an event by quieting, unquieting, or canceling it.
Procedure
1. Click Dashboard > Events > Summary.
2. Perform the following actions as needed.
l
To view additional information about an event, in the Actions column for that
event, click View details.
Procedure
1. Click Dashboard > Events > Event Notification Rules.
2. In the Notification Rules area on the Cluster Events page, click Add Rule.
3. In the Rule name field on the Add Notification Rule page, type a name for the rule.
4. In the Recipients area, specify a notification method.
a. To notify a recipient through email, select Email, type the email address to which
notifications will be sent, and then click Add.
b. To notify a community through SNMP traps, select SNMP, select the community
name and the SNMP host, which is the network monitoring station, from the
respective lists, and then click Add.
c. To add additional notification recipients or communities, repeat these steps.
5. In the Events area, expand the event types and select the check boxes for the events
and event types that you want to trigger this notification.
6. Click Submit.
55
Click Browse, navigate to and select the template file that you want to use, and
then click OK.
In the Set custom notification template field, type the path and file name of the
template file that you want to use.
5. Click Submit.
56
HEALTHY
Description
-
Interface
CLI, web
administration
interface
SMARTFAIL or
Smartfail or
restripe in
progress
NOT AVAILABLE
CLI, web
administration
interface
Error
state
Note
57
State
-
SUSPENDED
Description
-
Interface
-
CLI, web
administration
interface
NOT IN USE
REPLACE
CLI only
STALLED
CLI only
NEW
USED
CLI only
PREPARING
CLI only
EMPTY
CLI only
WRONG_TYPE
CLI only
BOOT_DRIVE
CLI only
SED_ERROR
CLI , web
administration
interface
Note
ERASE
58
CLI only
Error
state
State
-
INSECURE
Description
-
Interface
-
CLI only
Error
state
X
Note
UNENCRYPTED
SED
Web
administration
interface only
Note
SNMP monitoring
You can use SNMP to remotely monitor the EMC Isilon cluster hardware components,
such as fans, hardware sensors, power supplies, and disks. The default Linux SNMP tools
or a GUI-based SNMP tool of your choice can be used for this purpose.
You can enable SNMP monitoring on individual nodes on your cluster, and you can also
monitor cluster information from any node. Generated SNMP traps are sent to your SNMP
network. You can configure an event notification rule that specifies the network station
where you want to send SNMP traps for specific events, so that when an event occurs, the
cluster sends the trap to that server. OneFS supports SNMP in read-only mode. SNMP v1
and v2c is the default value, but you can configure settings for SNMP v3 alone or SNMP
v1, v2c, and v3.
Note
When SNMP v3 is used, OneFS requires the SNMP-specific security level of AuthNoPriv as
the default value when querying the cluster. The security level AuthPriv is not supported.
Elements in an SNMP hierarchy are arranged in a tree structure, similar to a directory tree.
As with directories, identifiers move from general to specific as the string progresses from
left to right. Unlike a file hierarchy, however, each element is not only named, but also
numbered.
For example, the SNMP
entity .iso.org.dod.internet.private.enterprises.isilon.oneFSss.s
sLocalNodeId.0 maps to .1.3.6.1.4.1.12124.3.2.0. The part of the name that
refers to the OneFS SNMP namespace is the 12124 element. Anything further to the right
of that number is related to OneFS-specific monitoring.
Management Information Base (MIB) documents define human-readable names for
managed objects and specify their data types and other properties. You can download
MIBs that are created for SNMP-monitoring of an Isilon cluster from the webadministration interface or manage them using the command-line interface. MIBs are
SNMP monitoring
59
stored in /usr/local/share/snmp/mibs/ on a OneFS node. The OneFS ISILONMIBs serve two purposes:
u
ISILON-MIB is a registered enterprise MIB. Isilon clusters have two separate MIBs:
ISILON-MIB
Defines a group of SNMP agents that respond to queries from a network monitoring
system (NMS) called OneFS Statistics Snapshot agents. As the name implies, these
agents snapshot the state of the OneFS file system at the time that it receives a
request and reports this information back to the NMS.
ISILON-TRAP-MIB
Generates SNMP traps to send to an SNMP monitoring station when the
circumstances occur that are defined in the trap protocol data units (PDUs).
The OneFS MIB files map the OneFS-specific object IDs with descriptions. Download or
copy MIB files to a directory where your SNMP tool can find them, such as /usr/share/
snmp/mibs/ or /usr/local/share/snmp/mibs, depending on the tool that you
use.
To have Net-SNMP tools read the MIBs to provide automatic name-to-OID mapping, add m All to the command, as in the following example.
snmpwalk -v2c -c public -m All <node IP> isilon
If the MIB files are not in the default Net-SNMP MIB directory, you may need to specify the
full path, as in the following example. Note that all three lines are one command.
snmpwalk -m /usr/local/share/snmp/mibs/ISILON-MIB.txt:/usr/local/
share/snmp/mibs/ISILON-TRAP-MIB.txt:/usr/local/share/snmp/mibs
/ONEFS-TRAP-MIB.txt \ -v2c -C c -c public <node IP> enterprises.onefs
Note
The examples are from running the snmpwalk command on a cluster. Your SNMP version
may require different arguments.
60
After you download and save the appropriate MIBs, you can configure SNMP monitoring
through either the web administration interface or though the command-line interface.
The Isilon cluster does not generate SNMP traps unless you configure an event
notification rule to send events.
Procedure
1. Click Cluster Management > General Settings > SNMP Monitoring.
2. In the Service area of the SNMP Monitoring page, enable or disable SNMP monitoring.
a. To disable SNMP monitoring, click Disable, and then click Submit.
b. To enable SNMP monitoring, click Enable, and then continue with the following
steps to configure your settings.
3. In the Downloads area, click Download for the MIB file that you want to download.
Follow the download process that is specific to your browser.
4. Optional: If you are using Internet Explorer as your browser, right-click the Download
link, select Save As from the menu, and save the file to your local drive.
You can save the text in the file format that is specific to your Net-SNMP tool.
5. Copy MIB files to a directory where your SNMP tool can find them, such as /usr/
share/snmp/mibs/ or /usr/local/share/snmp/mibs, depending on the
SNMP tool that you use.
To have Net-SNMP tools read the MIBs to provide automatic name-to-OID mapping,
add -m All to the command, as in the following example: snmpwalk -v2c -c
public -m All <node IP> isilon
6. Navigate back to the SNMP Monitoring page and configure General Settings.
a. In the Settings area, configure protocol access by selecting the version that you
want.
OneFS does not support writable OIDs; therefore, no write-only community string
setting is available.
b. In the System location field, type the system name.
This setting is the value that the node reports when responding to queries. Type a
name that helps to identify the location of the node.
c. Type the contact email address in the System contact field.
7. Optional: If you selected SNMP v1/v2 as your protocol, locate the SNMP v1/v2c
Settings section and type the community name in the Read-only community field.
SNMP monitoring
61
Cluster maintenance
Trained service personnel can replace or upgrade components in Isilon nodes.
Isilon Technical Support can assist you with replacing node components or upgrading
components to increase performance.
battery
SATA/SAS Drive
memory (DIMM)
fan
front panel
intrusion switch
IB/NVRAM card
SAS controller
NVRAM battery
power supply
If you configure your cluster to send alerts to Isilon, Isilon Technical Support will contact
you if a component needs to be replaced. If you do not configure your cluster to send
alerts to Isilon, you must initiate a service request.
62
drive
memory (DIMM)
If you want to upgrade components in your nodes, contact Isilon Technical Support.
63
Description
Shut down
Reboot
3. Click Submit.
The SupportIQ scripts are based on the Isilon isi_gather_info log-gathering tool.
The SupportIQ module is included with the OneFS operating system and does not require
you to activate a separate license. You must enable and configure the SupportIQ module
before SupportIQ can run scripts to gather data. The feature may have been enabled
when the cluster was first set up, but you can enable or disable SupportIQ through the
Isilon web administration interface.
In addition to enabling the SupportIQ module to allow the SupportIQ agent to run scripts,
you can enable remote access, which allows Isilon Technical Support personnel to
monitor cluster events and remotely manage your cluster using SSH or the web
administration interface. Remote access helps Isilon Technical Support to quickly identify
and troubleshoot cluster issues. Other diagnostic tools are available for you to use in
conjunction with Isilon Technical Support to gather and upload information such as
packet capture metrics.
Note
If you enable remote access, you must also share cluster login credentials with Isilon
Technical Support personnel. Isilon Technical Support personnel remotely access your
cluster only in the context of an open support case and only after receiving your
permission.
64
Configuring SupportIQ
OneFS logs contain data that Isilon Technical Support personnel can securely upload,
with your permission, and then analyze to troubleshoot cluster problems. The SupportIQ
technology must be enabled and configured for this process.
When SupportIQ is enabled, Isilon Technical Support personnel can request logs through
scripts that gather cluster data and then upload the data to a secure location. You must
enable and configure the SupportIQ module before SupportIQ can run scripts to gather
data. The feature may have been enabled when the cluster was first set up.
You can also enable remote access, which allows Isilon Technical Support personnel to
troubleshoot your cluster remotely and run additional data-gathering scripts. Remote
access is disabled by default. To enable remote SSH access to your cluster, you must
provide the cluster password to a Technical Support engineer.
Send alerts via SupportIQ agent (HTTPS) and by email (SMTP) SupportIQ
delivers notifications to Isilon through the SupportIQ agent over HTTPS and by
email over SMTP.
Configuring SupportIQ
65
Disable SupportIQ
You can disable SupportIQ so the SupportIQ agent does not run scripts to gather and
upload data about your EMC Isilon cluster.
Procedure
1. Click Cluster Management > General Settings > SupportIQ.
2. Clear the Enable SupportIQ check box.
3. Click Submit.
SupportIQ scripts
When SupportIQ is enabled, Isilon Technical Support personnel can request logs with
scripts that gather cluster data and then upload the data. The SupportIQ scripts are
located in the /usr/local/SupportIQ/Scripts/ directory on each node.
Data-gathering scripts
The following table lists the data-gathering activities that SupportIQ scripts perform.
These scripts can be run automatically, at the request of an Isilon Technical Support
representative, to collect information about your cluster's configuration settings and
operations. The SupportIQ agent then uploads the information to a secure Isilon FTP site,
so that it is available for Isilon Technical Support personnel to analyze. The SupportIQ
scripts do not affect cluster services or the availability of your data.
Action
-
66
Description
-
Action
-
Description
-
Collects and uploads information about the state and health of the
OneFS /ifs/ file system.
Get IB data
Collects and uploads only the most recent cluster log information.
Get messages
Collects and uploads information about cluster-wide and nodespecific network configuration settings and operations.
Warns if the chassis is open and uploads a text file of the event
information.
isi_gather_info
isi_gather_info -incremental
isi_gather_info
single node
SupportIQ scripts
67
Upgrading OneFS
Two options are available for upgrading the OneFS operating system: a rolling upgrade or
a simultaneous upgrade. Before upgrading OneFS softare, a pre-upgrade check must be
performed.
A rolling upgrade individually upgrades and restarts each node in the cluster
sequentially. During a rolling upgrade, the cluster remains online and continues serving
clients with no interruption in service, although some connection resets may occur on
SMB clients. Rolling upgrades are performed sequentially by node number, so a rolling
upgrade takes longer to complete than a simultaneous upgrade. The final node in the
upgrade process is the node that you used to start the upgrade process.
Note
Rolling upgrades are not available for all clusters. For instructions on how to upgrade the
cluster operating system, see the OneFS Release Notes.
A simultaneous upgrade installs the new operating system and restarts all nodes in the
cluster at the same time. Simultaneous upgrades are faster than rolling upgrades but
require a temporary interruption of service during the upgrade process. Your data is
inaccessible during the time that it takes to complete the upgrade process.
Before beginning either a simultaneous or rolling upgrade, OneFS compares the current
cluster and operating system with the new version to ensure that the cluster meets
certain criteria, such as configuration compatibility (SMB, LDAP, SmartPools), disk
availability, and the absence of critical cluster events. If upgrading puts the cluster at
risk, OneFS warns you, provides information about the risks, and prompts you to confirm
whether to continue the upgrade.
If the cluster does not meet the pre-upgrade criteria, the upgrade does not proceed, and
the unsupported statuses are listed.
68
CHAPTER 4
Access zones
Access zones
69
Access zones
Feature
Overlapping
share-name
support
Multiple access
zone support
Description
If multiple SMB share have the
same display name, OneFS
supports the overlapping display
names if the name appears only
once per access zone.
You can create additional access
zones and configure each zone
differently. Each access zone can
be configured with its own set of
authentication providers, user
mapping rules, and SMB shares.
Comment
For example, you can assign the name
"home" as the display name for a share
in zone A and a different share in zone
B.
Multiple access zones are particularly
useful for server consolidation, for
example when merging multiple
Windows file servers that are
potentially joined to different untrusted
forests.
Note
70
Access zones
Rules are called in the order they are listed. To ensure that each rule gets
processed, list replacements first and allow/deny rules last. You can change the
order in which a rule is listed by clicking its title bar and dragging it to a new
position.
6. Optional: From the SMB Shares list, select one of the following options:
7. Click Create Access Zone.
After you finish
Before you can use an access zone, you must associate it with an IP address pool.
Managing access zones
71
Access zones
Use all
authentication
providers
Manually select
authentication
providers
Description
-
2.
3.
4.
If you are finished adding provider instances, you can change the
priority in which they are called by changing the order in which
they are listed. To do so, click the title bar of a provider instance
and drag it up or down to a new position in the list.
Append fields from Modifies a token by adding specified fields to it. All appended
a user
identifiers become members of the additional groups list.
Insert fields from Modifies a token by adding specified fields from another token. An
a user
inserted primary user or group becomes the new primary user or group
in the token and moves the old primary user or group to the additional
identifiers list. Modifying the primary user leaves the tokens
username unchanged. When inserting additional groups from a token,
the new groups are added to the existing groups.
Replace a user
with a new user
Remove
Modifies a token by removing the supplemental groups.
supplemental
groups from a user
Join two users
together
Inserts the new token into the list of tokens. If the new token is the
second user, it is inserted after the existing token; otherwise, it is
inserted before the existing token. The insertion point is primarily
relevant when the existing token is already the first in the list because
the first token is used to determine the ownership of new system
objects.
Use all SMB shares Adds each available SMB share to the access zone.
Manually select
SMB shares
72
Allows you to select the SMB shares to add to the access zone. The
following steps are required:
Access zones
Setting
-
Description
1.
2.
In the Select SMB Shares dialog box, select the check box for
each SMB share that you want to add to the access zone.
3.
Click Select.
73
Access zones
74
CHAPTER 5
Authentication and access control
75
In most situations, the default settings are sufficient. You can configure additional access
zones, custom roles, and permissions policies as necessary for your particular
environment.
76
Feature
Access zones
Description
OneFS includes a built-in access zone
named System.
Comment
By default, new authentication
providers, SMB shares, and NFS
exports are added to the System
zone. When you create a new IP
address pool, you must assign it to
an access zone.
Authentication
Roles
Identity
management
Mixedenvironment
support
Feature
Description
(ACLs) and standard UNIX permissions
on the cluster file system.
Comment
identity mapping to translate
between Windows and UNIX
permissions as needed.
Although you can configure ACL policies to optimize a cluster for UNIX or Windows, you
should do so only if you understand how ACL and UNIX permissions interact.
The OneFS file system installs with UNIX permissions as the default. By using Windows
Explorer or OneFS administrative tools, you can give a file or directory an ACL. In addition
to Windows domain users and groups, ACLs in OneFS can include local, NIS, and LDAP
users and groups. After you give a file an ACL, OneFS stops enforcing the file's mode bits,
which remain only as an estimate of the effective permissions.
ACLs
In Windows environments, file and directory permissions, referred to as access rights, are
defined in access control lists (ACLs). Although ACLs are more complex than mode bits,
ACLs can express much more granular sets of access rules. OneFS uses the ACL
processing rules commonly associated with Windows ACLs.
A Windows ACL contains zero or more access control entries (ACEs), each of which
represents the security identifier (SID) of a user or a group as a trustee. In OneFS, an ACL
can contain ACEs with a UID, GID, or SID as the trustee. Each ACE contains a set of rights
that allow or deny access to a file or folder. An ACE can optionally contain an inheritance
flag to specify whether the ACE should be inherited by child folders and files.
Note
Instead of the standard three permissions available for mode bits, ACLs have 32 bits of
fine-grained access rights. Of these, the upper 16 bits are general and apply to all object
types. The lower 16 bits vary between files and directories but are defined in a way that
allows most applications to use the same bits for files and directories.
Data access control
77
Rights can be used for granting or denying access for a given trustee. A user's access can
be blocked explicitly through a deny ACE. Access can also be blocked implicitly by
ensuring that the user does not directly (or indirectly through a group) appear in an ACE
that grants the right in question.
UNIX permissions
In a UNIX environment, file and directory access is controlled by POSIX mode bits, which
grant read, write, or execute permissions to the owning user, the owning group, and
everyone else.
OneFS supports the standard UNIX tools for viewing and changing permissions, ls,
chmod, and chown. For more information, run the man ls, man chmod, and man
chown commands.
All files contain 16 permission bits, which provide information about the file or directory
type and the permissions. The lower 9 bits are grouped as three 3-bit sets, called triples,
which contain the read, write, and execute (rwx) permissions for each class of users
owner, group, and other. You can set permissions flags to grant permissions to each of
these classes.
Unless the user is root, OneFS uses the class to determine whether to grant or deny
access to the file. The classes are not cumulative; the first class matched is used. It is
therefore common to grant permissions in decreasing order.
Mixed-permission environments
When a file operation requests an objects authorization data (for example, with the ls
-l command over NFS or with the Security tab of the Properties dialog box in Windows
Explorer over SMB), OneFS attempts to provide that data in the requested format. In an
environment that mixes UNIX and Windows systems, some translation may be required
when performing create file, set security, get security, or access operations.
SID-to-UID and SID-to-GID mappings are cached in both the OneFS ID mapper and the
stat cache. If a mapping has recently changed, the file might report inaccurate
information until the file is updated or the cache is flushed.
each for owner, group, and everyone. For details about mode bits, see "UNIX
permissions."
2. Two access control entries (ACEs) are created for each triple: the allow ACE contains
the corresponding rights that are granted according to the permissions; the deny ACE
contains the corresponding rights that are denied. In both cases, the trustee of the
ACE corresponds to the file owner, group, or everyone. After all of the ACEs are
generated, any that are not needed are removed before the synthetic ACL is returned.
Roles
You can permit and limit access to administrative areas of your EMC Isilon cluster on a
per-user basis through the use of roles.
OneFS includes built-in administrator roles with predefined sets of privileges that cannot
be modified. The following list describes what you can and cannot do through roles:
u
You can add any user to a role as long as the user can authenticate to the cluster.
You can create custom roles and assign privileges to those roles.
You can add a group to a role, which grants to all users who are members of that
group all of the privileges associated with the role.
Note
When OneFS is first installed, only users with root- or admin-level can log in and assign
users to roles.
Built-in roles
Built-in roles include privileges to perform a set of administrative functions.
The following tables describe each of the built-in roles from most powerful to least
powerful. The tables include the privileges and read/write access levels (if applicable)
that are assigned to each role. You can assign users and groups to built-in roles and to
roles that you create.
Table 1 SecurityAdmin role
Description
-
Privileges
-
Read/write
access
ISI_PRIV_LOGIN_CONSOLE N/A
ISI_PRIV_LOGIN_PAPI
N/A
ISI_PRIV_LOGIN_SSH
N/A
ISI_PRIV_AUTH
Read/write
ISI_PRIV_ROLE
Read/write
Roles
79
Description
-
Privileges
-
ISI_PRIV_LOGIN_CONSOLE
Read/write
access
N/A
ISI_PRIV_LOGIN_PAPI
N/A
ISI_PRIV_LOGIN_SSH
N/A
ISI_PRIV_SYS_SHUTDOWN
N/A
ISI_PRIV_SYS_SUPPORT
N/A
ISI_PRIV_SYS_TIME
N/A
ISI_PRIV_ANTIVIRUS
Read/write
ISI_PRIV_AUDIT
Read/write
ISI_PRIV_CLUSTER
Read/write
ISI_PRIV_DEVICES
Read/write
ISI_PRIV_EVENT
Read/write
ISI_PRIV_FTP
Read/write
ISI_PRIV_HTTP
Read/write
ISI_PRIV_ISCSI
Read/write
ISI_PRIV_JOB_ENGINE
Read/write
ISI_PRIV_LICENSE
Read/write
ISI_PRIV_NDMP
Read/write
ISI_PRIV_NETWORK
Read/write
ISI_PRIV_NFS
Read/write
ISI_PRIV_NTP
Read/write
ISI_PRIV_QUOTA
Read/write
ISI_PRIV_REMOTE_SUPPORT Read/write
80
ISI_PRIV_SMARTPOOLS
Read/write
ISI_PRIV_SMB
Read/write
ISI_PRIV_SNAPSHOT
Read/write
ISI_PRIV_STATISTICS
Read/write
ISI_PRIV_SYNCIQ
Read/write
ISI_PRIV_VCENTER
Read/write
ISI_PRIV_NS_TRAVERSE
N/A
ISI_PRIV_NS_IFS_ACCESS
N/A
Description
Privileges
Read/write access
N/A
ISI_PRIV_LOGIN_PAPI
N/A
ISI_PRIV_LOGIN_SSH
N/A
ISI_PRIV_ANTIVIRUS
Read-only
ISI_PRIV_AUDIT
Read-only
ISI_PRIV_CLUSTER
Read-only
ISI_PRIV_DEVICES
Read-only
ISI_PRIV_EVENT
Read-only
ISI_PRIV_FTP
Read-only
ISI_PRIV_HTTP
Read-only
ISI_PRIV_ISCSI
Read-only
ISI_PRIV_JOB_ENGINE
Read-only
ISI_PRIV_LICENSE
Read-only
SI_PRIV_NDMP
Read-only
ISI_PRIV_NETWORK
Read-only
ISI_PRIV_NFS
Read-only
ISI_PRIV_NTP
Read-only
ISI_PRIV_QUOTA
Read-only
ISI_PRIV_REMOTE_SUPPORT Read-only
ISI_PRIV_SMARTPOOLS
Read-only
ISI_PRIV_SMB
Read-only
ISI_PRIV_SNAPSHOT
Read-only
ISI_PRIV_STATISTICS
Read-only
ISI_PRIV_SYNCIQ
Read-only
ISI_PRIV_VCENTER
Read-only
Description
-
Privileges
-
ISI_PRIV_LOGIN_PAPI
Read/write
access
N/A
ISI_PRIV_ISCSI
Read/write
ISI_PRIV_NETWORK
Read/write
ISI_PRIV_SMARTPOOLS
Read/write
Built-in roles
81
Description
Privileges
ISI_PRIV_SNAPSHOT
Read/write
access
Read/write
ISI_PRIV_SYNCIQ
Read/write
ISI_PRIV_VCENTER
Read/write
ISI_PRIV_NS_TRAVERSE
N/A
ISI_PRIV_NS_IFS_ACCESS N/A
OneFS privileges
Privileges in OneFS are assigned through role membership; privileges cannot be assigned
directly to users and groups.
Table 5 Login privileges
OneFS privilege
User right
Privilege type
Action
ISI_PRIV_LOGIN_PAPI
Action
ISI_PRIV_LOGIN_SSH
Action
OneFS privilege
User right
Privilege type
Action
ISI_PRIV_SYS_SUPPORT
Action
ISI_PRIV_SYS_TIME
Action
OneFS privilege
ISI_PRIV_AUTH
ISI_PRIV_ROLE
82
User right
Configure external
authentication providers
Create new roles and
assign privileges
Privilege type
Read/write
Read/write
OneFS privilege
ISI_PRIV_ANTIVIRUS
User right
Configure antivirus
scanning
Privilege type
Read/write
IS_PRIV_AUDIT
Configure audit
capabilities
Read/write
ISI_PRIV_CLUSTER
Read/write
ISI_PRIV_DEVICES
Read/write
ISI_PRIV_EVENT
Read/write
ISI_PRIV_FTP
Read/write
ISI_PRIV_HTTP
Read/write
ISI_PRIV_ISCSI
Read/write
ISI_PRIV_JOB_ENGINE
Schedule cluster-wide
jobs
Read/write
ISI_PRIV_LICENSE
Read/write
ISI_PRIV_NDMP
Read/write
ISI_PRIV_NETWORK
Configure network
interfaces
Read/write
ISI_PRIV_NFS
Read/write
ISI_PRIV_NTP
Configure NTP
Read/write
ISI_PRIV_QUOTA
Read/write
ISI_PRIV_REMOTE_SUPPO
RT
Read/write
ISI_PRIV_SMARTPOOLS
Read/write
ISI_PRIV_SMB
Read/write
ISI_PRIV_SNAPSHOT
Read/write
ISI_PRIV_SNMP
Read/write
ISI_PRIV_STATISTICS
Read/write
ISI_PRIV_SYNCIQ
Configure SyncIQ
Read/write
ISI_PRIV_VCENTER
Read/write
OneFS privileges
83
OneFS privilege
ISI_PRIV_NS_TRAVERSE
ISI_PRIV_NS_IFS_ACCESS
User right
Traverse and view
directory metadata
Privilege type
Action
OneFS privilege
ISI_PRIV_EVENT
User right
View and modify system
events
Privilege type
Read/write
ISI_PRIV_LICENSE
Read/write
ISI_PRIV_STATISTICS
Read/write
However, if you are on the sudoers list, the following command succeeds:
sudo isi sync policy list
The following tables list all One FS commands available, the associated privilege or rootaccess requirement, and whether sudo is required to run the command.
Note
If you are running in compliance mode, additional sudo commands are available.
Table 11 Privileges sorted by CLI command
isi command
isi alert
isi audit
84
Privilege
ISI_PRIV_EVENT
ISI_PRIV_AUDIT
Requires sudo
x
isi command
isi auth - excluding isi auth
role
Privilege
ISI_PRIV_AUTH
Requires sudo
ISI_PRIV_ROLE
isi avscan
ISI_PRIV_ANTIVIRUS
isi batterystatus
ISI_PRIV_STATISTICS
isi config
root
ISI_PRIV_JOB_ENGINE
ISI_PRIV_STATISTICS
isi devices
ISI_PRIV_DEVICES
isi domain
root
isi email
ISI_PRIV_CLUSTER
isi events
ISI_PRIV_EVENT
isi exttools
root
isi fc
root
isi filepool
ISI_PRIV_SMARTPOOLS
isi firmware
root
isi ftp
ISI_PRIV_FTP
isi get
root
isi hdfs
root
isi iscsi
ISI_PRIV_ISCSI
isi job
ISI_PRIV_JOB_ENGINE
isi license
ISI_PRIV_LICENSE
isi lun
ISI_PRIV_ISCSI
isi ndmp
ISI_PRIV_NDMP
isi networks
ISI_PRIV_NETWORK
isi nfs
ISI_PRIV_NFS
isi perfstat
ISI_PRIV_STATISTICS
isi pkg
root
isi quota
ISI_PRIV_QUOTA
isi readonly
root
85
isi command
isi remotesupport
Privilege
ISI_PRIV_REMOTE_SUPPORT
Requires sudo
isi servicelight
ISI_PRIV_DEVICES
isi services
root
isi set
root
isi smartlock
root
isi smb
ISI_PRIV_SMB
isi snapshot
ISI_PRIV_SNAPSHOT
isi snmp
ISI_PRIV_SNMP
isi stat
ISI_PRIV_STATISTICS
isi statistics
ISI_PRIV_STATISTICS
isi status
ISI_PRIV_STATISTICS
isi storagepool
ISI_PRIV_SMARTPOOLS
isi sync
ISI_PRIV_SYNCIQ
isi tape
ISI_PRIV_NDMP
isi target
ISI_PRIV_ISCSI
isi update
root
isi version
ISI_PRIV_CLUSTER
isi worm
root
isi zone
ISI_PRIV_AUTH
Privilege
ISI_PRIV_ANTIVIRUS
isi commands
isi avscan
ISI_PRIV_AUDIT
isi audit
ISI_PRIV_AUTH
Requires sudo
x
isi zone
ISI_PRIV_CLUSTER
isi email
isi version
ISI_PRIV_DEVICES
isi devices
isi servicelight
86
Privilege
ISI_PRIV_EVENT
isi commands
isi alert
Requires sudo
x
isi events
ISI_PRIV_FTP
isi ftp
ISI_PRIV_ISCSI
isi iscsi
isi lun
isi target
ISI_PRIV_JOB_ENGINE
isi job
isi dedupe - excluding isi dedupe
stats
ISI_PRIV_LICENSE
isi license
ISI_PRIV_NDMP
isi ndmp
isi tape
ISI_PRIV_NETWORK
isi networks
ISI_PRIV_NFS
isi nfs
ISI_PRIV_QUOTA
isi quota
ISI_PRIV_ROLE
ISI_PRIV_REMOTE_SUPPORT
isi remotesupport
ISI_PRIV_SMARTPOOLS
isi filepool
isi storagepool
ISI_PRIV_SMB
isi smb
ISI_PRIV_SNAPSHOT
isi snapshot
ISI_PRIV_SNMP
isi snmp
ISI_PRIV_STATISTICS
isi batterystatus
isi sync
root
isi config
isi domain
isi exttools
isi fc
Command-line interface privileges
87
Privilege
isi commands
isi firmware
Requires sudo
isi get
isi hdfs
isi pkg
isi readonly
isi services
isi set
isi smartlock
isi update
isi worm
Authentication
OneFS supports local and remote authentication providers to verify that users attempting
to access the cluster are who they claim to be. Anonymous access, which does not
require authentication, is supported for protocols that allow it.
OneFS supports the concurrent use of multiple authentication provider types, which are
analogous to directory services. For example, OneFS is often configured to authenticate
Windows clients with Active Directory and to authenticate UNIX clients with LDAP. It is
important that you understand their interactions before enabling multiple providers on
the cluster.
Note
88
Authentic
ation
provider
-
Active
Directory
LDAP
NTLM
x
*
Kerberos
User/
Netgroups UNIX
Windows
group
properties properties
managem
ent
-
NIS
Local
File
x
x
Feature
Authentication
Description
All authentication providers support plain-text authentication.
Some providers can be configured to support NTLM or Kerberos
authentication also.
Netgroups
LDAP
The Lightweight Directory Access Protocol (LDAP) is a networking protocol that enables
you to define, query, and modify directory services and resources.
OneFS can authenticate users and groups against an LDAP repository in order to grant
them access to the cluster. OneFS supports Kerberos authentication for an LDAP provider.
The LDAP service supports the following features:
u
Configurable LDAP schemas. For example, the ldapsam schema allows NTLM
authentication over the SMB protocol for users with Windows-like attributes.
Redundancy and load balancing across servers with identical directory data.
Multiple LDAP provider instances for accessing servers with different user data.
Encrypted passwords.
Authentication provider features
89
Active Directory
The Active Directory directory service is a Microsoft implementation of Lightweight
Directory Access Protocol (LDAP), Kerberos, and DNS technologies that can store
information about network resources. Active Directory can serve many functions, but the
primary reason for joining the cluster to an Active Directory domain is to perform user and
group authentication.
When the cluster joins an Active Directory domain, a single Active Directory machine
account is created. The machine account is used to establish a trust relationship with the
domain and to enable the cluster to authenticate and authorize users in the Active
Directory forest. By default, the machine account is named the same as the cluster;
however, if the cluster name is more than 15 characters long, the name is hashed and
displayed after joining the domain.
Whenever possible, a single Active Directory instance should be used when all domains
have a trust relationship. Multiple instances should be used only to grant access to
multiple sets of mutually-untrusted domains.
Note
NIS
The Network Information Service (NIS) provides authentication and identity uniformity
across local area networks. OneFS includes a NIS authentication provider that enables
you to integrate the cluster with your NIS infrastructure.
NIS, designed by Sun Microsystems, can be used to authenticate users and groups when
they access the cluster. The NIS provider exposes the passwd, group, and netgroup maps
from a NIS server. Hostname lookups are also supported. Multiple servers can be
specified for redundancy and load balancing.
Note
File provider
A file provider enables you to supply an authoritative third-party source of user and group
information to the cluster. A third-party source is useful in UNIX and Linux environments
that synchronize /etc/passwd, /etc/group, and etc/netgroup files across
multiple servers.
OneFS uses standard BSD /etc/spwd.db and /etc/group database files as the
backing store for the file provider. You generate the spwd.db file by running the
pwd_mkdb command in the OneFS command-line interface (CLI). You can script updates
to the database files.
On the Isilon cluster, a file provider uses libcrypt for password hashing. The Modular
Crypt Format is parsed to determine the hashing algorithm. The following algorithms are
supported:
90
MD5
Blowfish
NT-Hash
SHA-256
SHA-512
Note
The built-in System file provider includes services to list, manage, and authenticate
against system accounts such as root, admin, and nobody. It is recommended that you
do not modify the System file provider.
Local provider
The local provider provides authentication and lookup facilities for user accounts that
were added by an administrator.
Local authentication can be useful when Active Directory, LDAP, or NIS directory services
are not used, or when a specific user or application needs to access the cluster. Local
groups can include built-in groups and Active Directory groups as members.
In addition to configuring network-based authentication sources, you can also manage
local users and groups by configuring a local password policy for each node in the
cluster. OneFS settings specify password complexity, password age and re-use, and
password-attempt lockout policies.
l
l
native: Let OneFS determine the identity to store on disk. This is the
recommended setting.
unix: Always store incoming UNIX identifiers (UIDs and GIDs) on disk.
sid: Store incoming Windows security identifiers (SIDs) on disk, unless the
SID was generated from a UNIX identifier; in that case, convert it back to the
UNIX identifier and store it on disk.
Local provider
91
Workgroup
Specifies the NetBIOS workgroup. The default value is WORKGROUP.
Space Replacement
For clients that have difficulty parsing spaces in user and group names, specifies
a substitute character.
3. Click Save.
After you finish
If you changed the on-disk identity selection, it is recommended that you run the Repair
Permissions job with the 'Convert permissions' repair task to prevent potential
permissions errors.
Because ACL policies change the behavior of permissions throughout the system, they
should be modified only as necessary by experienced administrators with advanced
knowledge of Windows ACLs. This is especially true for the advanced settings, which are
applied regardless of the cluster's environment.
For UNIX, Windows, or balanced environments, the optimal permission policy settings are
selected and cannot be modified. However, you can choose to manually configure the
cluster's default permission settings if necessary to support your particular environment.
Note
You must be logged in to the web administration interface to perform this task.
For a description of each setting option, see ACL policy settings options.
Procedure
1. Click Protocols > ACLs > ACL Policies.
2. In the Standard Settings section, under Environment, click to select the setting that
best describes your environment, or select Configure permission policies manually to
configure individual permission policies.
3. If you selected the Configure permission policies manually option, configure the
settings as needed.
For more information about these settings, see ACL policy settings options.
4. In the Advanced Settings section, configure the settings as needed.
92
Setting
UNIX only
Description
Causes cluster permissions to operate with UNIX semantics, as opposed to
Windows semantics. Enabling this option prevents ACL creation on the system.
Setting
Balanced
Description
Causes cluster permissions to operate in a mixed UNIX and Windows
environment. This setting is recommended for most cluster deployments.
Windows only
Configure
permission
policies
manually
ACL creation
over SMB
Specifies whether to allow or deny creation of ACLs over SMB. Select one of the
following options.
l
Note
Inheritable ACLs on the system take precedence over this setting: If inheritable
ACLs are set on a folder, any new files and folders created in that folder will
inherit the folder's ACL. Disabling this setting does not remove ACLs currently
set on files. If you want to clear an existing ACL, run the chmod -b <mode>
<file> command to remove the ACL and set the correct permissions.
chmod on files Controls what happens when a chmod operation is initiated on a file with an
ACL, either locally or over NFS. This setting controls any elements that set UNIX
with existing
permissions, including File System Explorer. Enabling this policy setting does
ACLs
not change how chmod operations affect files that do not have ACLs. Select
one of the following options.
l
Remove the existing ACL and set UNIX permissions instead: For
chmod operations, removes any existing ACL and instead sets the chmod
permissions. Select this option only if you do not need permissions to be
set from Windows.
Remove the existing ACL and create an ACL equivalent to the UNIX
permissions: Stores the UNIX permissions in a Windows ACL. Select this
option only if you want to remove Windows permissions but do not want
files to have synthetic ACLs.
Remove the existing ACL and create an ACL equivalent to the UNIX
permissions, for all users/groups referenced in old ACL: Stores the
UNIX permissions in a Windows ACL. Select this option only if you want to
remove Windows permissions but do not want files to have synthetic ACLs.
Merge the new permissions with the existing ACL: Causes Windows
and UNIX permissions to operate smoothly in a balanced environment by
merging permissions that are applied by chmod with existing ACLs. An
ACE for each identity (owner, group, and everyone) is either modified or
created, but all other ACEs are unmodified. Inheritable ACEs are also left
unmodified to enable Windows users to continue to inherit appropriate
93
Setting
Description
permissions. However, UNIX users can set specific permissions for each of
those three standard identities.
l
Deny permission to modify the ACL: Prevents users from making NFS
and local chmod operations. Enable this setting if you do not want to allow
permission sets over NFS. This setting returns an error when an NFS client
attempts to modify the ACL.
If you try to run the chmod command on the same permissions that are
currently set on a file with an ACL, you may cause the operation to silently fail
The operation appears to be successful, but if you were to examine the
permissions on the cluster, you would notice that the chmod command had
no effect. As a workaround, you can run the chmod command away from the
current permissions and then perform a second chmod command to revert to
the original permissions. For example, if your file shows 755 UNIX
permissions and you want to confirm this number, you could run chmod 700
file; chmod 755 file
ACLs created
on directories
by UNIX
chmod
On Windows systems, the access control entries for directories can define finegrained rules for inheritance; on UNIX, the mode bits are not inherited. Making
ACLs that are created on directories by the chmod command inheritable is
more secure for tightly controlled environments but may deny access to some
Windows users who would otherwise expect access.
Select one of the following options.
l
chown on files Changes a file or folder's owning user or group. Select one of the following
options.
with existing
ACLs
l
Modify the owner and/or group permissions: Causes the chown
operation to perform as it does in UNIX. Enabling this setting modifies any
ACEs in the ACL associated with the old and new owner or group.
l
Do not modify the ACL: Cause the NFS chown operation to function as it
does in Windows. When a file owner is changed over Windows, no
permissions in the ACL are changed.
94
Setting
Description
Note
Over NFS, the chown operation changes the permissions and the owner or
owning group. For example, consider a file owned by user Joe with "rwx------"
(700) permissions, signifying "rwx" permissions for the owner, but no
permissions for anyone else. If you run the chown command to change
ownership of the file to user Bob, the owner permissions are still "rwx" but they
now represent the permissions for Bob, rather than for Joe. In fact, Joe will have
lost all of his permissions. This setting does not affect UNIX chown operations
performed on files with UNIX permissions, and it does not affect Windows
chown operations, which do not change any permissions.
Access checks In UNIX environments, only the file owner or superuser has the right to run a
chmod or chown operation on a file. In Windows environments, you can
(chmod,
implement this policy setting to give users the right to perform chmod
chown)
operations, called the "change permissions" right, or the right to perform
chown operations, called the "take ownership" right.
Note
The "take ownership" right only gives users the ability to take file ownership,
not to give ownership away.
Select one of the following options.
l
Allow owner and users with 'take ownership' right to chown, and
owner and users with 'change permissions' right to chmod: Causes
chmod and chown access checks to operate with Windows-like behavior.
Treatment of
"rwx"
permissions
95
Setting
Description
l
Group owner
inheritance
Operating systems tend to work with group ownership and permissions in two
different ways: BSD inherits the group owner from the file's parent folder;
Windows and Linux inherit the group owner from the file creator's primary
group. If you enable a setting that causes the group owner to be inherited from
the creator's primary group, it can be overridden on a per-folder basis by
running the chmod command to set the set-gid bit. This inheritance applies
only when the file is created. For more information, see the manual page for
the chmod command.
Select one of the following options.
l
BSD semantics - Inherit group owner from the parent folder: Causes
the group owner to be inherited from the file's parent folder.
chmod (007)
on files with
existing ACLs
Specifies whether to remove ACLs when running the chmod (007) command.
Select one of the following options.
l
Owner
permissions
96
Setting
Description
l
Approximate owner mode bits using only the ACE with the owner
ID: Makes the owner permissions appear more accurate, in that you see
only the permissions for a particular owner and not the more permissive
set. However, this may cause access-denied problems for UNIX clients.
group
permissions
Approximate group mode bits using all possible group ACEs: Makes
the group permissions appear more permissive than the actual
permissions on the file.
Approximate group mode bits using only the ACE with the group
ID: Makes the group permissions appear more accurate, in that you see
only the permissions for a particular group and not the more permissive
set. However, this may cause access-denied problems for UNIX clients.
No "deny"
ACEs
The Windows ACL user interface cannot display an ACL if any "deny" ACEs are
out of canonical ACL order. However, in order to correctly represent UNIX
permissions, deny ACEs may be required to be out of canonical ACL order.
Select one of the following options.
l
Remove deny ACEs from synthetic ACLs: Does not include "deny"
ACEs when generating synthetic ACLs. This setting can cause ACLs to be
more permissive than the equivalent mode bits.
Access check
(utimes)
You can control who can change utimes, which are the access and
modification times of a file, by selecting one of the following options.
l
97
job with the 'convert permissions' task to ensure that the changes are fully propagated
throughout the cluster.
Procedure
1. Click Protocols > ACLs > Repair Permissions Job.
2. Optional: From the Priority list, select the priority level at which to run the job in
relation to other jobs.
3. Optional: From the Impact policy list, select an impact policy for the job to follow.
4. From the Repair task list, select one of the following methods for updating
permissions:
Options
Description
Convert
permissions
Clone
permissions
Inherit
permissions
Description
8. Optional: From the Access Zone list, select an access zone to use for ID mapping. This
setting applies to only the Convert permissions repair task.
98
Managing roles
You can view, add, or remove members of any role. Except for built-in roles, whose
privileges you cannot modify, you can add or remove OneFS privileges on a role-by-role
basis.
Note
Roles take both users and groups as members. If a group is added to a role, all users who
are members of that group are assigned the privileges associated with the role. Similarly,
members of multiple roles are assigned the combined privileges of each role.
View roles
You can view information about built-in and custom roles.
This procedure must be performed through the command-line interface (CLI).
Procedure
1. Establish an SSH connection to any node in the cluster.
2. At the command prompt, run one of the following commands.
l
To view detailed information about each role on the cluster, including member and
privilege lists, run:
isi auth roles list --verbose
To view detailed information about a single role, run the following command,
where <role> is the name of the role:
isi auth roles view <role>
View privileges
You can view user privileges.
This procedure must be performed through the command-line interface (CLI). You can
view a list of your privileges or the privileges of another user using the following
commands:
Procedure
1. At the command prompt, run one of the following commands.
l
To view a list of privileges for another user, where <user> specifies the user by
name:
isi auth mapping token <user>
Managing roles
99
Modify a role
You can modify the description and the user or group membership of any role, including
built-in roles. However, you cannot modify the name or privileges that are assigned to
built-in roles.
This procedure must be performed through the command-line interface (CLI).
Procedure
1. Establish an SSH connection to any node in the cluster.
2. At the command prompt, run the following command, where <role> is the role name
and <options> are optional parameters:
isi auth roles modify <role> [<options>]
100
authenticate to it. You also can create a password policy for each local provider to
enforce password complexity.
If you do not specify a port, the default port is used. The default port for non-secure
LDAP (ldap://) is 389; for secure LDAP (ldaps://) it is 636. If you specify non-secure
LDAP, the bind password is transmitted to the server in clear text.
5. Optional: Configure the following settings as needed.
Load balance servers
Select the check box to connect to a random server, or clear the check box to
connect according to the order in which the servers are listed in the Servers field.
Base Distinguished Name
Type the distinguished name (DN) of the entry at which to start LDAP searches.
Base DNs can include cn (Common Name), l (Locality), dc (Domain Component),
ou (Organizational Unit), or other components. For example, dc=emc,dc=com is a
base DN for emc.com.
Bind to
Type the distinguished name of the entry at which to bind to the LDAP server.
Password
Specify the password to use when binding to the LDAP server. Use of this
password does not require a secure connection; if the connection is not using
Transport Layer Security (TLS), the password is sent in clear text.
6. Optional: To modify the default settings for user, group, and netgroup queries, click
Default Query Settings.
7. Optional: To modify the settings for user queries and home directory provisioning,
click User Query Settings.
8. Optional: To modify the settings for group queries, click Group Query Settings.
9. Optional: To modify the settings for netgroup queries, click Netgroup Query Settings.
10.Optional: To modify the default LDAP attributes that contain user information or to
modify LDAP security settings, click Advanced LDAP Settings.
11.Click Add LDAP provider.
101
Most settings require modification only if you are using a Kerberos Key Distribution
Center (KDC) other than Active Directoryfor example, if you are using an MIT KDC for NFS
version 3 or version 4 authentication.
Procedure
1. Establish an SSH connection to any node in the cluster.
2. Run the isi auth krb5 command with the add, modify, or delete subcommand to specify which entries to modify in the Kerberos configuration file.
3. Write the changes to the file by running the isi auth krb5 write command.
Note
By default, changes are written to the global Kerberos configuration file, /etc/
krb5.conf. To update the Kerberos configuration file for Active Directory, include
the --path option to specify the /etc/likewise-krb5-ad.conf file.
102
103
104
If you migrate users to a new or different Active Directory domain, you must re-set the ACL
domain information after you configure the new provider. Third-party tools can be used,
such as Microsoft SubInACL.
Procedure
1. Click Cluster Management > Access Management > Active Directory.
2. Click Join a domain.
3. In the Domain Name field, type a fully qualified Active Directory domain name.
The domain name will also be used as the provider name.
4. In the User field, type the username of an account that is authorized to join the Active
Directory domain.
5. In the Password field, type the password of the user account.
6. Optional: In the Organizational Unit field, type the name of the organizational unit
(OU) to connect to on the Active Directory server. Specify the OU in the form OuName or
OuName1/SubName2.
7. Optional: In the Machine Account field, type the name of the machine account.
105
Note
If you specified an OU to connect to, the domain join will fail if the machine account
does not reside in the OU.
8. Optional: To enable Active Directory authentication for NFS, select the Enable Secure
NFS check box.
If you enable this setting, OneFS registers NFS service principal names (SPNs) during
the domain join.
9. Optional: To configure advanced settings, click Advanced Active Directory Settings.
10.Click Join.
106
Note
Most settings require modification only if you are using a Kerberos Key Distribution
Center (KDC) other than Active Directoryfor example, if you are using an MIT KDC for NFS
version 3 or version 4 authentication.
Procedure
1. Establish an SSH connection to any node in the cluster.
2. Run the isi auth krb5 command with the add, modify, or delete subcommand to specify which entries to modify in the Kerberos configuration file.
3. Write the changes to the file by running the isi auth krb5 write command.
Note
By default, changes are written to the global Kerberos configuration file, /etc/
krb5.conf. To update the Kerberos configuration file for Active Directory, include
the --path option to specify the /etc/likewise-krb5-ad.conf file.
Description
-
Trusted Domains
Domains to Ignore
Enhanced Privacy
Creates a home directory the first time a user logs in, if a home
directory does not already exist for the user.
UNIX Shell
Specifies the path to the login shell to use if the Active Directory
server does not provide login-shell information. This setting applies
only to users who access the file system through SSH.
107
Setting
-
Lookup User
Description
-
Auto-assign UIDs
Lookup Group
Auto-assign GIDs
Make UID/GID
Restricts user and group lookups to the specified domains.
assignments for users and
groups in these specific
domains
If the Load balance servers option is not selected, servers are accessed in the order in
which they are listed.
5. Optional: Configure the Load balance servers setting:
l
To connect according to the order in which the servers are listed in the Servers
field, clear the check box.
6. Optional: Click Default Query Settings and then configure the following settings:
NIS Domain
Specifies the NIS domain name.
108
Search Timeout
Specifies the number of seconds after which to stop retrying and fail a search.
The default value is 100.
Retry Frequency
Specifies the timeout period in seconds after which a request will be retried. The
default value is 5.
7. Optional: Click User Query Settings and then configure the following settings:
Authenticate users from this provider
Specifies whether to allow the provider to respond to authentication requests.
Home Directory Naming
Specifies the path to use as a template for naming home directories. The path
must begin with /ifs and can contain variables, such as %U, that are expanded
to generate the home directory path for the user.
Create home directories on first login
Specifies whether to create a home directory the first time a user logs in, if a
home directory does not already exist for the user.
UNIX Shell
Specifies the path to the user's login shell, for users who access the file system
through SSH.
8. Optional: Click Host Name Query Settings and then configure the Resolve Hosts
setting:
l
109
To specify a user replacement file, in the Users File field, type or browse to the
location of the spwd.db file.
To specify a group replacement file, in the Groups File field, type or browse to the
location of the group file.
To specify a netgroup replacement file, in the Netgroups File field, type or browse
to the location of the netgroup file.
5. Optional: To enable this provider to authenticate users, select the Authenticate users
from this provider check box.
6. Optional: In the Home Directory Naming field, type the path to use as a template for
naming home directories. The path must begin with /ifs and may contain variables,
such as %U, that are expanded to generate the home directory path for the user.
7. Optional: To automatically create a home directory the first time a user logs in, select
the Create home directories on first login check box.
This setting applies only if a home directory does not already exist for the user.
8. Optional: From the UNIX Shell list, select the login shell for users who access the file
system through SSH.
9. Click Add File Provider.
110
Note
If you omit the -d option, the file is created in the /etc directory. For full command
usage guidelines, view the manual ("man") page by running the man pwd_mkdb
command.
The following command generates an spwd.db file in the /ifs directory from a
password file that is located at /ifs/test.passwd:
pwd_mkdb -d /ifs /ifs/test.passwd
If the replacement files are located outside the /ifs directory tree, you must manually
distribute them to every node in the cluster. Changes that are made to the system
provider's files are automatically distributed across the cluster.
111
The fields are defined below in the order in which they appear in the file.
Note
UNIX systems often define the passwd format as a subset of these fields, omitting the
Class, Change, and Expiry fields. To convert a file from passwd to master.passwd
format, add :0:0: between the GID field and the Gecos field.
Username
The user name. This field is case-sensitive. OneFS does not set a limit on the length;
however, many applications truncate the name to 16 characters.
Password
The users encrypted password. If authentication is not required for the user, an
asterisk (*) can be substituted for a password. The asterisk character is guaranteed
to not match any password.
UID
The UNIX user identifier. This value must be a number in the range 0-4294967294
that is not reserved or already assigned to a user. Compatibility issues will occur if
this value conflicts with an existing account's UID.
GID
The group identifier of the users primary group. All users are a member of at least
one group, which is used for access checks and can also be used when creating
files.
Class
This field is not supported by OneFS and should be left empty.
Change
OneFS does not support changing the passwords of users in the file provider. This
field is ignored.
Expiry
OneFS does not support the expiration of user accounts in the file provider. This field
is ignored.
Gecos
This field can store a variety of information but is usually used to store the users full
name.
Home
The absolute path to the users home directory, beginning at /ifs.
112
Shell
The absolute path to the users shell. If this field is set to /sbin/nologin, the user
is denied command-line access.
The fields are defined below in the order in which they appear in the file.
Group name
The name of the group. This field is case-sensitive. Although OneFS does not set a
limit on the length of the group name, many applications truncate the name to 16
characters.
Password
This field is not supported by OneFS and should contain an asterisk (*).
GID
The UNIX group identifier. Valid values are any number in the range 0-4294967294
that is not reserved or already assigned to a group. Compatibility issues will occur if
this value conflicts with an existing group's GID.
Group members
A comma-delimited list of user names.
A new line signifies a new netgroup. You can continue a long netgroup entry to the next
line by typing a backslash character (\) in the right-most position of the first line.
113
Allow password to expire: Select this check box to specify that the password is
allowed to expire.
UID: If this setting is left blank, the system automatically allocates a UID for the
account. This is the recommended setting.
Note
You cannot assign a UID that is in use by another local user account.
l
Additional Groups: Specify any additional groups to make this user a member of.
Home Directory: Type the path to the user's home directory. If you do not specify a
path, a directory is automatically created at /ifs/home/<Username>.
UNIX Shell: This setting applies only to users who access the file system through
SSH. From the list, click the shell that you want. By default, the /bin/zsh shell is
selected.
Enabled: Select this check box to allow the user to authenticate against the local
database for SSH, FTP, HTTP, and Windows file sharing through SMB. This setting
is not used for UNIX file sharing through NFS.
Prompt password change: Select this check box to prompt for a password change
the next time the user logs in.
114
You cannot assign a GID that is in use by another group. It is recommended that you
leave this field blank to allow the system to automatically generate the GID.
7. Optional: Follow these steps for each member that you want to add the group:
a. For the Members setting, click Add user. The Select a User dialog box appears.
b. For the Search for setting, select either Users or Well-known SIDs.
c. If you selected Users, specify values for the following fields:
Username: Type all or part of a user name, or leave the field blank to return all
users. Wildcard characters are accepted.
Access Zone: Select the access zone that contains the authentication provider
that you want to search.
Provider: Select an authentication provider.
d. Click Search.
e. In the Search Results table, select a user and then click Select.
The dialog box closes.
8. Click Create.
115
Option
Description
Users
Groups
3. From the Select a zone list, select an access zone (for example, System).
4. From the Select a provider list, select the local provider for the access zone (for
example, LOCAL:System).
116
You must configure a separate password policy for each access zone. Each access zone
in the cluster contains a separate instance of the local provider, which allows each
access zone to have its own list of local users who can authenticate. Password
complexity is configured for each local provider, not for each user.
Procedure
1. Establish an SSH connection to any node in the cluster.
2. Optional: To view the current password settings, run the following command:
isi auth local view system
3. Run the isi auth local modify command, choosing from local password policy
settings options.
The --password-complexity parameter must be specified for each setting, as
demonstrated in the following example:
isi auth local modify system --password-complexity=lowercase \
--password-complexity=uppercase --password-complexity=numeric \
--password-complexity=symbol
The following command is an example of how to configure a local password policy for a
local provider.
isi auth local modify provider-name=<provider-name> \
--min-password-length=15 \
--lockout-duration=15m \
--lockout-window=1m \
--lockout-threshold=5 \
--add-password-complexity=uppercase \
--add-password-complexity=numeric
117
number of possible passwords that an attacker must check before the correct password
can be guessed.
-
Setting
min-password-length
password-complexity
Description
Minimum password length
in characters.
Comments
Long passwords are best. The
minimum length should not be so
long that users have a difficult time
entering or remembering the
password.
You can specify as many as four
cases. The following cases are valid:
l
uppercase
lowercase
numeric
min-password-age
max-password-age
password-historylength
lockout-duration
lockout-threshold
118
Setting
lockout-window
Description
Comments
119
CHAPTER 6
Identity management
Identity management
121
Identity management
Identity management
In environments with several different types of directory services, OneFS maps the users
and groups from the separate services to provide a single unified identity on the EMC
Isilon cluster and uniform access control to files and directories, regardless of the
incoming protocol. This process is called identity mapping.
Isilon clusters are frequently deployed in multiprotocol environments with multiple types
of directory services, such as Active Directory and LDAP. When a user with accounts in
multiple directory services logs in to an Isilon cluster, OneFS combines the users
identities and privileges from all the directory services into a native access token. You
can configure OneFS settings to include a list of rules for token manipulation to control
user identity and privileges. For example, you can set a user mapping rule to merge an
Active Directory identity and an LDAP identity into a single token that works for access to
files stored over both SMB and NFS. The token can include groups from Active Directory
and LDAP. The mapping rules that you create can solve identity problems by
manipulating access tokens in many ways, including the following examples:
u
Authenticate a user with Active Directory but give the user a UNIX identity.
Disallow login of users that do not exist in both Active Directory and LDAP.
For more information about identity management, see the white paper Managing
identities with the Isilon OneFS user mapping service (white paper) at EMC Online
Support (https://support.emc.com).
Identity types
OneFS supports three primary identity types, each of which can be stored directly on the
file system. These types are user identifier and group identifier for UNIX, and security
identifier for Windows.
When you log on to an Isilon cluster, the user mapper expands your identity to include
your other identities from all the directory services, including Active Directory, LDAP, and
NIS. After OneFS maps your identities across the directory services, it generates an
access token that includes the identity information associated with your accounts. A
token includes the following identifiers:
u
A UNIX user identifier (UID) and a group identifier (GID). A UID or GID is a 32-bit
number with a maximum value of 4,294,967,295.
A security identifier (SID) for a Windows user account. A SID is a series of authorities
and sub-authorities ending with a 32-bit relative identifier (RID). Most SIDs have the
form S-1-5-21-A-B-C-<RID> , where A, B, and C are specific to a domain or
computer and <RID> denotes the object in the domain.
A list of supplemental identities, including all groups in which the user is a member.
The token also contains privileges that stem from administrative role-based access
control.
On an Isilon cluster, a file contains permissions, which appear as an access control list
(ACL). The ACL controls access to directories, files, and other securable system objects.
When a user tries to access a file, OneFS compares the identities in the users access
token with the files ACL. OneFS grants access when the files ACL includes an access
control entry (ACE) that allows the identity in the token to access the file and that does
122
Identity management
not include an ACE that denies the identity access. OneFS compares the access token of
a user with the ACL of a file.
Note
For more information about access control lists, including a description of the
permissions and how they correspond to POSIX mode bits, see the white paper titled
EMC Isilon multiprotocol data access with a unified security model on the EMC Online
Support web site (https://support.emc.com).
When a name is provided as an identifier, it is converted into the corresponding user or
group object and the correct identity type. There are various ways that a name can be
entered or displayed:
u
UNIX assumes unique case-sensitive namespaces for users and groups. For example,
"Name" and "name" represent different objects.
Windows provides a single, case-insensitive namespace for all objects and also
specifies a prefix to target an Active Directory domain (for example, domain\name).
Kerberos and NFSv4 define principals, which require names to be formatted the same
way as email addresses (for example, name@domain.com).
Multiple names can reference the same object. For example, given the name "support"
and the domain "example.com", support, EXAMPLE\support, and support@example.com
are all names for a single object in Active Directory.
Access tokens
An access token is created when the user first makes a request for access.
Access tokens represent who a user is when performing actions on the cluster and supply
the primary owner and group identities to use during file creation. Access tokens are also
compared against the ACL or mode bits during authorization checks.
During user authorization, OneFS compares the access token, which is generated during
the initial connection, with the authorization data on the file. All user and identity
mapping occurs during token generation; no mapping takes place during permissions
evaluation.
An access token includes all UIDs, GIDs, and SIDs for an identity, in addition to all OneFS
privileges. OneFS exclusively uses the information in the token to determine whether a
user has access to a resource. It is important that the token contains the correct list of
UIDs, GIDs, and SIDs.
An access token is created from one of the following sources:
Source
-
Username
Authorization method
-
123
Identity management
Source
-
Authorization method
-
ID mapping
The ID mapping service maps Windows SIDs to UNIX UIDs and, conversely, to control
access consistently across protocols. Administrators with advanced knowledge of UNIX
and Windows identities can modify the default settings that determine how identities are
mapped in the system.
Note
Identity (ID) mapping and user mapping are different services, despite the similarity in
names.
During authentication, the ID mapping service associates Windows identifiers with UNIX
identifiers. When a user connects to a cluster over NFS, the ID mapping service maps the
users UID and GID to a SID for access to files that another user stored over SMB. In the
same way, when a user connects to the cluster over SMB with a SID, the service maps it
to a UID and GID for access to files stored over NFS by a UNIX client. By default, the ID
mapping service matches accounts with the same name.
Mappings are stored in a cluster-distributed database called the ID mapper. When
retrieving a mapping from the database, the ID mapper takes a source and target identity
type as input. If a mapping already exists between the specified source and the
requested type, that mapping is returned; otherwise, a new mapping is created. Each
mapping is stored in the ID mapper database as a one-way relationship from the source
to the target identity type. Two-way mappings are stored as complementary one-way
mappings.
124
Identity management
The user or group is in the primary domain or one of the listed lookup domains.
For UIDs, generate a UNIX SID with a domain of S-1-22-1 and a resource ID (RID)
matching the UID. For example, the UNIX SID for UID 600 is S-1-22-1-600.
For GIDs, generate a UNIX SID with a domain of S-1-22-2 and a RID matching the
GID. For example, the UNIX SID for GID 800 is S-1-22-2-800.
125
Identity management
Note
You can configure mapping rules when you create an access zone. OneFS maps users
only during login or protocol access.
If you do not configure rules, a user who authenticates with one directory service receives
full access to the identity information in other directory services when the account names
are the same. For example, a user who authenticates with an Active Directory domain as
Desktop\jane automatically receives permissions for the corresponding UNIX user
account for jane from LDAP or NIS.
In the most common scenario, OneFS is connected to two directory services, Active
Directory and LDAP. In such a case, the default mapping provides a user with a UID from
LDAP and a SID from the default group in Active Directory. The user's groups come from
Active Directory and LDAP, with the LDAP groups added to the list. To pull groups from
LDAP, the mapping service queries the memberUid. The users home directory, gecos,
and shell come from Active Directory.
126
Options
A parameter
Wildcards
Identity management
Operator
append
Web interface
Append fields
from a user
CLI Direction
++
Left-to-right
Description
Modifies an access token by adding fields to
it. The mapping service appends the fields
that are specified in the list of options (user,
group, groups) to the first identity in the
rule. The fields are copied from the second
identity in the rule. All appended identifiers
become members of the additional groups
list. An append rule without an option
performs only a lookup operation; you must
include an option to alter a token.
insert
Insert fields
from a user
+=
Left-to-right
replace
Replace one
user with a
different user
=>
Left-to-right
user error.
remove
groups
Remove
supplemental
groups from a
user
--
Unary
127
Identity management
Operator
join
Web interface
Join two users
together
CLI Direction
-
Description
&= Bidirectional Inserts the new identity into the token. If the
new identity is the second user, the
mapping service inserts it after the existing
identity; otherwise, the service inserts it
before the existing identity. The location of
the insertion point is relevant when the
existing identity is already the first in the list
because OneFS uses the first identity to
determine the ownership of new file system
objects.
username
unix_name
primary_uid
primary_user_sid
primary_gid
primary_group_sid
Options control how a rule combines identity information in a token. The break option is
the exception: It stops OneFS from processing additional rules.
Although several options can apply to a rule, not all options apply to all operators. The
following table describes the effect of each option and the operators that they work with.
-
128
Option
user
Operator
insert, append
Description
Copies the primary UID
and primary user SID, if
they exist, to the token.
groups
insert, append
groups
insert, append
Identity management
Option
Operator
Description
exclude the primary UID,
the primary GID, the
primary user SID, and the
primary group SID.)
default_user
break
all operators
Best practice
Use Active Directory with
RFC 2307 and Windows
Services for UNIX
Comments
Use Microsoft Active Directory with Windows Services for UNIX and
RFC 2307 attributes to manage Linux, UNIX, and Windows systems.
Integrating UNIX and Linux systems with Active Directory centralizes
identity management and eases interoperability, reducing the need
for user mapping rules. Make sure your domain controllers are
running Windows Server 2003 or later.
Employ a consistent
username strategy
129
Identity management
Best practice
Avoid common UIDs and
GIDs
Comments
You should not use well-known UIDs and GIDs in your ID ranges
because they are reserved for system accounts. UIDs and GIDs below
1000 are reserved for system accounts; do not assign them to users
or groups.
1.
Place the rules that replace an identity first to ensure that OneFS
replaces all instances of the identity.
2.
3.
4.
Within each group of rules, put explicit rules before rules with
wildcards; otherwise, the explicit rules might be skipped.
On-disk identity
After the user mapper resolves a user's identities, OneFS determines an authoritative
identifier for it, which is the preferred on-disk identity.
OnesFS stores either UNIX or Windows identities in file metadata on disk. On-disk identity
types are UNIX, SID, and native. Identities are set when a file is created or a file's access
control data is modified. Almost all protocols require some level of mapping to operate
correctly, so choosing the preferred identity to store on disk is important. You can
configure OneFS to store either the UNIX or the Windows identity, or you can allow OneFS
to determine the optimal identity to store.
On-disk identity types are UNIX, SID, and native. Although you can change the type of ondisk identity, the native identity is best for a network with UNIX and Windows systems. In
native mode, setting the UID as the on-disk identity improves NFS performance.
Note
When you upgrade from a version of OneFS that is older than 7.0, the on-disk identity is
set to UNIX. The SID on-disk identity is for a homogeneous network of Windows systems
managed only with Active Directory. On new installations, the on-disk identity is set to
native.
130
Identity management
The native on-disk identity type allows the OneFS authentication daemon to select the
correct identity to store on disk by checking for the identity mapping types in the
following order:
Order Mapping
type
-
Description
-
Algorithmic
mapping
External
mapping
A user with an explicit UID and GID defined in a directory service (such as
Active Directory with RFC 2307 attributes, LDAP, NIS, or the OneFS file
provider or local provider) has the UNIX identity set as the on-disk
identity.
Persistent
mapping
No mapping
If a user lacks a UID or GID even after querying the other directory
services and identity databases, its SID is set as the on-disk identity. In
addition, to make sure a user can access files over NFS, OneFS allocates
a UID and GID from a preset range of 1,000,000 to 2,000,000. In native
mode, a UID or GID that OneFS generates is never set as the on-disk
identity.
Note
If you change the on-disk identity, you should run the repairpermissions job.
On-disk identity
131
CHAPTER 7
Auditing
Auditing overview................................................................................................134
Protocol audit events.......................................................................................... 134
Supported event types........................................................................................ 134
Supported audit tools......................................................................................... 135
Enable system configuration auditing..................................................................136
Enable protocol access auditing..........................................................................136
Auditing settings................................................................................................. 137
Integrating with the EMC Common Event Enabler.................................................137
Auditing
133
Auditing
Auditing overview
You can audit system configuration and SMB protocol activity on the Isilon cluster. All
audit data is stored and protected in the cluster file system and organized in files called
audit topics.
You can export SMB audit data to Varonis DatAdvantage or other third-party vendors that
support the EMC Common Event Enabler (CEE) framework. You can view system
configuration activity on each node through a command-line tool.
You can enable or disable system configuration auditing; no additional configuration is
required. If you enable configuration auditing, all configuration events that are handled
by the APIincluding writes, modifications, and deletionsare tracked and recorded in
the config audit topic.
You can enable and configure protocol auditing for one or more access zones in the Isilon
cluster. If you enable protocol auditing for an access zone, file-access events through the
SMB protocol are recorded in the protocol audit topic. The protocol audit topic is
consumable by auditing applications that support the EMC Common Event Enabler (CEE),
such as Varonis DatAdvantage for Windows. By default, OneFS logs only the events that
are handled by Varonis, but you can specify which events to log in each access zone. For
example, you might want to audit the default set of protocol events in the System access
zone but audit only successful attempts to delete files in a different access zone.
Auditing
create
delete
set_security
The following event types are available for forwarding through CEE but are unsupported
by Varonis DatAdvantage:
-
read
write
close
get_security
The following protocol audit events are not exported through CEE and are unsupported by
Varonis DatAdvantage:
-
logon
logoff
It is recommended that you install and configure third-party auditing applications before
you enable the OneFS auditing feature. Otherwise, the backlog consumed by the tool may
be so large that results may be stale for a prolonged time.
-
Application
Supported features
Audit events
create
rename
135
Auditing
Application
Supported features
Audit events
set_security
It is recommended that you install and configure third-party auditing applications before
you enable the OneFS auditing feature. Otherwise, the backlog consumed by the tool may
be so large that results may be stale for a prolonged time.
Procedure
1. Click Cluster Management > Auditing.
2. In the Settings area, select the Enable Configuration Change Auditing checkbox.
3. Click Save Changes.
It is recommended that you install and configure third-party auditing applications before
you enable the OneFS auditing feature. Otherwise, the backlog consumed by the tool may
be so large that results may be stale for a prolonged time.
Procedure
1. Click Cluster Management > Auditing.
2. In the Settings area, select the Enable Protocol Access Auditing checkbox.
3. In the Audited Zones area, click Add Zones.
4. In the Select Zones dialog box, select the checkbox for one or more access zones, and
then click Add Zones.
5. Optional: In the Event Forwarding area, specify one or more CEE servers to forward
logged events to.
a. In the CEE Server URIs field, type the URI of each CEE server in the CEE server pool.
The OneFS CEE export service uses round robin load-balancing when exporting
events to multiple CEE servers. Valid URIs start with http:// and include the port
number and path to the CEE server if necessary.
http://example.com:12228/cee
b. In the Storage Cluster Name field, specify the name of the storage cluster to use
when forwarding protocol events.
136
Auditing
This value is typically the SmartConnect zone name. This setting is required only if
needed by your third-party audit application.
6. Click Save Changes.
Results
The following protocol events, which are the only events supported by Varonis
DatAdvantage, are collected for audited access zones by default: create, delete,
rename, and set_security. You can modify the set of events that are audited in an
access zone by running the isi zone zones modify command in the command-line
interface. Because each audited event consumes system resources, it is recommended
that you only configure zones for events that are needed by your auditing application.
Auditing settings
You can view or modify basic settings for configuration change auditing and protocol
access auditing.
Enable Configuration Change Auditing
Audits requests that are made through the API for system configuration changes.
Enable Protocol Access Auditing
Audits requests that are made through the SMB protocol to access data.
Audited Zones
Specifies one or more access zones to audit. This setting applies only to protocol
access auditing.
CEE Server URIs
Specifies one or more CEE server URIs where audit events will be forwarded. The
OneFS CEE export service uses round robin load-balancing when exporting events to
multiple CEE servers. This setting applies only to protocol access auditing.
Storage Cluster Name
Specifies the name of the storage cluster to use when forwarding protocol events
typically, the SmartConnect zone name. This setting is required only if needed by
your third-party audit application.
It is recommended that you install and configure third-party auditing applications before
you enable the OneFS auditing feature. Otherwise, the backlog consumed by the tool may
be so large that results may be stale for a prolonged time.
Auditing settings
137
Auditing
Install WinRAR or another suitable archival program that can open .iso files as an
archive, and copy the files.
Install SlySoft Virtual CloneDrive, which allows you to mount an ISO image as a drive
that you can copy files from.
Note
138
Auditing
Setting
CEE HTTP
listen port
Registry location
[HKEY_LOCAL_MACHINE\SOFTWARE\EMC\CEE
\Configuration]
Key
HttpPort
Value
12228
Enable
audit
remote
endpoints
[HKEY_LOCAL_MACHINE\SOFTWARE\EMC\CEE\CEPP
\Audit\Configuration]
Enabled
Audit
remote
endpoints
[HKEY_LOCAL_MACHINE\SOFTWARE\EMC\CEE\CEPP
\Audit\Configuration]
EndPoint
<EndPoint>
Note
l
The HttpPort value must match the port in the CEE URIs that you specify during
OneFS protocol audit configuration.
139
CHAPTER 8
File sharing
File sharing
141
File sharing
It is recommended that you do not save data to the root /ifs file path but in directories
below /ifs. The design of your data storage structure should be planned carefully. A
well-designed directory optimizes cluster performance and cluster administration.
You can set Windows- and UNIX-based permissions on OneFS files and directories. Users
who have the required permissions and administrative privileges can create, modify, and
read data on the cluster through one or more of the supported file sharing protocols.
u
SMB. Allows Microsoft Windows and Mac OS X clients to access files that are stored
on the cluster.
NFS. Allows UNIX, Linux, Mac OS X, Solaris, and other UNIX-based clients to access
files that are stored on the cluster.
HTTP and HTTPS (with optional DAV). Allows clients to access files that are stored on
the cluster through a web browser.
FTP. Allows any client that is equipped with an FTP client program to access files that
are stored on the cluster through the FTP protocol.
SMB
OneFS includes a configurable SMB service to create and manage SMB shares. SMB
shares provide Windows clients network access to file system resources on the cluster.
You can grant permissions to users and groups to carry out operations such as reading,
writing, and setting access permissions on SMB shares.
The /ifs directory is configured as an SMB share and is enabled by default. OneFS
supports both user and anonymous security modes. If the user security mode is enabled,
users who connect to a share from an SMB client must provide a valid user name with
proper credentials.
The SMB protocol uses security identifiers (SIDs) for authorization data. All identities are
converted to SIDs during retrieval and are converted back to their on-disk representation
before they are stored on the cluster.
When a file or directory is created, OneFS checks the access control list (ACL) of its parent
directory. If the ACL contains any inheritable access control entries (ACEs), a new ACL is
generated from those ACEs. Otherwise, OneFS creates an ACL from the combined file and
directory create mask and create mode settings.
OneFS supports the following SMB clients:
142
File sharing
2.1
Windows 7 or later
Windows Server 2008 R2 or later
NFS
OneFS includes a configurable NFS service to create and manage NFS exports, which
provide UNIX clients network access to file system resources on the cluster.
The Isilon cluster supports NIS and LDAP authentication providers for NFS file sharing.
OneFS supports asynchronous and synchronous communication over NFS.
143
File sharing
FTP
OneFS includes a secure FTP service called vsftpd, which stands for Very Secure FTP
Daemon, that you can configure for standard FTP and FTPS file transfers.
It is recommended that you configure ACL and UNIX permissions only if you fully
understand how they interact with one another.
It is recommended that you keep write caching enabled. You should also enable write
caching for all file pool policies.
OneFS interprets writes to the cluster as either synchronous or asynchronous, depending
on a client's specifications. The impacts and risks of write caching depend on what
protocols clients use to write to the cluster, and whether the writes are interpreted as
synchronous or asynchronous. If you disable write caching, client specifications are
ignored and all writes are performed synchronously.
The following table explains how clients' specifications are interpreted, according to the
protocol.
144
File sharing
Protocol Synchronous
-
Asynchronous
-
NFS
SMB
iSCSI
Protocol Risk
-
NFS
If a node fails, no data will be lost except in the unlikely event that a client of that
node also crashes before it can reconnect to the cluster. In that situation,
asynchronous writes that have not been committed to disk will be lost.
SMB
If a node fails, asynchronous writes that have not been committed to disk will be lost.
iSCSI
CAUTION
If a node fails, asynchronous writes that have not been committed can cause
inconsistencies in any file system that is laid out on the LUN, rendering the file
system unusable.
It is recommended that you do not disable write caching, regardless of the protocol that
you are writing with. If you are writing to the cluster with asynchronous writes, and you
decide that the risks of data loss are too great, it is recommended that you configure your
clients to use synchronous writes, rather than disable write caching.
145
File sharing
Modifying the advanced settings could result in operational failures. Be aware of the
potential consequences before committing changes to these settings.
Procedure
1. Click Protocols > Windows Sharing (SMB) > SMB Settings.
2. For the SMB service setting, select Enabled.
3. To configure advanced SMB server settings, click SMB Server Settings.
4. To configure advanced SMB share settings, click SMB Share Settings.
5. Click Save.
Full Control
Read-Write
Read
5. Click Save.
146
File sharing
These settings affect the behavior of the SMB service. Changes to these settings can
affect all current and future SMB shares.
Setting
-
Setting value
-
Visible at Root
Accessible at Root
Visible in Subdirectories
Accessible in
Subdirectories
Changes that are made from the SMB Settings tab override the default settings for all
SMB shares.
If the mask and mode bits match the default values, a green check mark next to a setting
appears, indicating that the specified read (R), write (W), or execute (X) permission is
enabled at the user, group, or "other" level. The "other" level includes all users who are
not listed as the owner of the share, and are not part of the group level that the file
belongs to.
Setting
-
Setting value
-
Create Permissions Sets the default source permissions to apply when a file or directory is
created. The default value is Default ACL.
Create Mask (Dir)
Specifies UNIX mode bits that are removed when a directory is created,
restricting permissions. Mask bits are applied before mode bits are applied.
Specifies UNIX mode bits that are added when a directory is created,
enabling permissions. Mode bits are applied after mask bits are applied.
Specifies UNIX mode bits that are removed when a file is created, restricting
permissions. Mask bits are applied before mode bits are applied.
Specifies UNIX mode bits that are added when a file is created, enabling
permissions. Mode bits are applied after mask bits are applied.
147
File sharing
Changes that are made from the SMB Settings tab override the default settings for all
SMB shares.
Setting
-
Setting value
-
Change Notify Configures notification of clients when files or directories change. This helps
prevent clients from seeing stale content, but requires server resources. The
default value is Norecurse.
Oplocks
Changes that are made from the SMB Settings tab override the default settings for all
SMB shares.
Setting
-
Setting value
-
Impersonate Guest Determines guest access to a share. The default value is Never.
Impersonate User
Allows all file access to be performed as a specific user. This must be a fully
qualified user name. The default value is No value.
NTFS ACL
Allows ACLs to be stored and edited from SMB clients. The default value is
Yes.
148
File sharing
Note
It is recommended that you configure advanced SMB share settings only if you have a
solid understanding of the SMB protocol.
You can specify one or more of the following variables in the directory path but you
must select the Allow Variable Expansion check box or the string is interpreted
literally by the system.
Variable Expansion
-
%D
%U
%Z
%L
%0
%1
%2
For example, if a user is in a domain named DOMAIN and has a username of user_1,
the path /ifs/home/%D/%U expands to /ifs/home/DOMAIN/user_1.
6. Apply the initial ACL settings for the directory. You can modify these settings later.
l
To apply a default ACL to the shared directory, click Apply Windows default ACLs.
149
File sharing
Note
If the Auto-Create Directories setting is enabled, OneFS creates an ACL with the
equivalent of UNIX 700 mode bit permissions for any directory that is created
automatically.
l
To maintain the existing permissions on the shared directory, click Do not change
existing permissions.
To expand path variables such as %U in the share directory path, select Allow
Variable Expansion.
To automatically create home directories when users access the share for the first
time, select Auto-Create Directories. This option is available only if the Allow
Variable Expansion option is enabled.
SMB shares are currently created with read-only permissions by default. To enable access
to a share, you must modify the share settings to allow users to write to the share. This
functionality is available only through the OneFS command-line interface. For example,
the following command allows the well-known user Everyone full permissions to a share
named HOMEDIR:
isi smb shares permission modify HOMEDIR --wellknown
--permission-type allow --permission full
Everyone \
Any changes made to these settings will only affect the settings for this share. If you need
to make changes to the global default values, that can be done from the SMB Settings
tab.
Procedure
1. Click Protocols > Windows Sharing (SMB) > SMB Shares.
2. From the list of SMB shares, locate the share you want to modify and then click View
details.
3. For each setting that you want to modify, click Edit, make the change, and then click
Save.
4. To modify the settings for file and directory permissions, performance, or security,
click Advanced SMB Share Settings.
150
File sharing
Description
Users
Enter the username you want to search for in the text field, and
then click Search.
Groups
Enter the group you want to search for in the text field, and then
click Search.
151
File sharing
6. Select the SMB share you want to add to this zone and click Select.
7. In the SMB Shares in this Zone list, click Edit to modify the share.
8. Type the overlapping display name you want in the Display Name field and click Save.
9. In the SMB Shares area, click Save.
10.Repeat this procedure for each SMB share that will use the same display name.
After you finish
You must associate an IP address pool with each access zone. The IP address the user
connects through specifies which zone the user is allowed to access.
You can delete all of the shares on the cluster by selecting the Name/Path option, and
then selecting Delete from the drop-down menu.
3. Click Delete.
4. In the confirmation dialog box, click Delete to confirm the deletion.
File sharing
the settings for individual NFS exports as you create them, or edit the settings for
individual exports as needed.
Procedure
1. Click Protocols > UNIX Sharing (NFS) > NFS Settings.
2. Enable or disable the NFS service and version support settings:
l
NFS Service
NFSv2 Support
NFSv3 Support
NFSv4 Support
Description
153
File sharing
5. Click Save.
Service
Description
-
NFSv2 support
NFSv3 support
NFSv4 support
Lock protection level Determines the number of node failures that can happen before a lock may
be lost. The default value is +2.
Block size
Description
-
The block size reported to NFSv2+ clients. The default value is 8192.
Commit asynchronously If set to yes, allows NFSv3 and NFSv4 COMMIT operations to be
asynchronous. The default value is No.
154
The preferred directory read transfer size reported to NFSv3 and NFSv4
clients. The default value is 131072.
The maximum read transfer size reported to NFSv3 and NFSv4 clients.
The default value is 1048576.
The preferred read transfer size reported to NFSv3 and NFSv4 clients.
Readdirplus prefetch
Setattr asynchronous
The reply to send for DATASYNC writes. The default value is DATASYNC.
File sharing
Setting
-
Description
-
The reply to send for FILESYNC writes. The default value is FILESYNC.
The maximum write transfer size reported to NFSv3 and NFSv4 clients.
The default value is 1048576.
The preferred write transfer size reported to NFSv3 and NFSv4 clients.
The default value is 524288.
The reply to send for UNSTABLE writes. The default value is UNSTABLE.
Setting value
-
Readdirplus enable
Setting value
-
Permits non-root users to set file times. The default value is Yes.
Overrides the general encoding settings the cluster has for the export. The
default value is DEFAULT.
Map Lookup UID Looks up incoming user identifiers (UIDs) in the local authentication database.
The default value is No.
Symlinks
Enables symlink support for the export. The default value is Yes.
Time delta
155
File sharing
Changes to the advanced settings affect all current and future NFS exports that use
default settings, and may impact the availability of the NFS file sharing service. Do not
make changes to these settings unless you have experience working with NFS. It is
recommended that you change the default values for individual NFS exports as you create
them, or edit the settings of existing exports.
Modifying the global default values is not recommended. You can override the settings
for NFS exports as you create them, or modify the settings for existing exports.
Procedure
1. Click Protocols > UNIX Sharing (NFS) > NFS Settings.
2. Click the NFS export settings menu.
3. For each setting that you want to modify, click System Default in the list of options
and select Custom Default.
Note
File sharing
If no clients are specified, all clients are allowed to access the export. If you add the
same client to more than one list and the client is entered in the same format for each
entry, the client is normalized to a single list in the following order of priority: Root
Clients, Always Read-Write Clients, Always Read-Only Clients,
Clients.
Setting
Description
Clients
Always
Read-Write
Clients
Always
Read-Only
Clients
Root Clients
5. For the Directory Paths setting, type or browse to the directory that you want to export.
You can add more directory paths by clicking Add another directory path.
6. Specify export permissions:
l
UNIX (system)
Kerberos5
Kerberos5 Integrity
Kerberos5 Privacy
Create an NFS export
157
File sharing
You can delete all the exports on a cluster by selecting the Export ID/Path option, and
then selecting Delete from the drop-down menu.
Procedure
1. Click Protocols > UNIX Sharing (NFS) > NFS Export
2. From the list of NFS exports, click the check box for the export that you want to delete.
3. Click Delete.
4. In the confirmation dialog box, click Delete to confirm the deletion.
In the following example output, export 1 contains a path that does not currently
exist:
ID
Message
----------------------------------1
'/ifs/test' is not a directory
----------------------------------Total: 1
158
File sharing
Description
Server-toserver
transfers
Anonymous
access
Local access
4. Click Submit.
Description
Enable HTTP
159
File sharing
Option
Description
Disable HTTP entirely Closes the HTTP port used for file access. Users can continue
to access the web administration interface by specifying the
port number in the URL. The default port is 8080.
3. In the Document root directory field, type or click Browse to navigate to an existing
directory in /ifs, or click File System Explorer to create a new directory and set its
permissions.
Note
The HTTP server runs as the daemon user and group. To properly enforce access
controls, you must grant the daemon user or group read access to all files under the
document root, and allow the HTTP server to traverse the document root.
4. In the Server hostname field, type the HTTP server name. The server hostname must
be a fully-qualified, SmartConnect zone name and valid DNS name. The name must
begin with a letter and contain only letters, numbers, and hyphens (-).
5. In the Administrator email address field, type an email address to display as the
primary contact for issues that occur while serving files.
6. From the Active Directory Authentication list, select an authentication setting:
Option
Description
Off
Integrated Authentication
Only
7. Click the Enable DAV check box. This allows multiple users to manage and modify files
collaboratively across remote web servers.
8. Click the Disable access logging check box.
9. Click Submit.
Home directories
When you create a local user, OneFS automatically creates a home directory for the user.
OneFS also supports dynamic home directory provisioning for users who access the
cluster by connecting to an SMB share or by logging in through FTP or SSH. Regardless of
160
File sharing
the method by which a home directory was created, you can configure access to the
home directory through a combination of SMB, SSH, and FTP.
Because SMB sends an NT password hash to authenticate SMB users, only users from
authentication providers that can handle NT hashes can log in over SMB. These providers
include the local provider, Active Directory, and LDAP with Samba extensions enabled.
File, NIS, and non-Samba LDAP users cannot log in over SMB.
Share permissions are checked when files are accessed, before the underlying file
system permissions are checked. Either of these permissions can prevent access to the
file or directory.
Home directory share paths must begin with /ifs/ and must be within the root path of
the access zone in which the home directory SMB share is created.
161
File sharing
# cd /ifs/home/user411
# ls -lde .
drwx------ + 2 user411 <your-company> Users 0 Oct 19 16:23 ./
OWNER: user:user411
GROUP: group:<your-company> Users
CONTROL:dacl_auto_inherited,dacl_protected
0: user:user411 allow dir_gen_all,object_inherit,container_inherit
162
File sharing
In this example from a Windows client, when the net use command is run on m:, Zachary
sees the contents of his /ifs/home/zachary directory:
# net use m: \\cluster.ip\zachary /u:zachary # cd m: # dir
In this example, another user, Claudia, sees the directory contents of /ifs/home/
claudia:
# net use m: \\cluster.ip\claudia /u:claudia # cd m: # dir
If Claudia tries to access Zachary's share, however, she cannot connect because it does
not exist for her.
Note
If another SMB share exists that matches the user's name, then the user connects to the
explicitly named share rather than the %U share.
Active Directory
Local
Note
File, NIS, and non-SAM LDAP providers cannot handle NT hashes, so users from those
authentication providers cannot log in over SMB.
163
File sharing
Create Home Directory. A boolean value that indicates whether to create the
home directory if it does not exist.
Home Directory Template. The template path name for the user's home
directory, and may contain special variables beginning with '%' that are expanded to
generate the home directory path for the user. The path name must begin with /
ifs/.
Login Shell. The default login shell for the user. The user's login shell may also
be provided by the authentication provider.
Note
A user must have the ISI_PRIV_LOGIN_SSH privilege to log in to a node through SSH.
The following example demonstrates setting these options for an Active Directory
authentication provider.
# id YOUR\\user_100
uid=1000008(<your-domain>\user_100) gid=1000000(<your-domain>\domain
users) \
groups=1000000(<your-domain>\domain users),1000024(<your-domain>\c1t),
1545(Users)
The information is verified from an external Unix node, as seen in the result of the
following command:
# ssh <your-domain>\\user_100@cluster.isilon.com
164
File sharing
This example sets the default login shell for all Active Directory users in <your-domain>
to /bin/bash:
You can modify the umask option for a zone with the --home-directory-umask option,
specifying an octal number as the umask.
The following example demonstrates how to allow a group/others write/execute
permission in a home directory. In this example, the user's home directory is created with
mode bits 0755 masked by the umask field, set to the value of 022. So by default, a
user's home directory is created with mode bits 0700 (equivalent to (0755 & ~(077)) :
165
File sharing
166
File sharing
This command reveals the path set in the Skeleton Directory field of the user's
access zone:
Local
Home directory
creation
-
/ifs/home/%U
Enabled
UNIX login
shell
-
/bin/sh
File
None
Disabled
None
Active Directory
/ifs/home/%D/%U
Disabled
/bin/sh
Note
If available, provider
information overrides this
value.
LDAP
None
Disabled
None
NIS
None
Disabled
None
When you create an SMB share through the web administration interface, you must select
the Allow Variable Expansion check box or the string is interpreted literally by the system.
-
Variable Value
-
%U
%D
Description
Expands to the user name to allow different users
to use different home directories. This variable is
typically included at the end of the path. For
example, for a user named user1, the path /ifs/
home/%U is mapped to /ifs/home/user1.
Expands to the user's domain name, based on the
authentication provider:
167
File sharing
Variable Value
-
Description
l
%Z
%L
%0
%1
%2
Note
If the user name includes fewer than three characters, the %0, %1, and %2 variables
wrap around. For example, for a user named ab, the variable maps to a, b, and a. For a
user named a, all three variables map to a.
168
Local User
File sharing
File User
LDAP User
NIS User
169
CHAPTER 9
Snapshots
Snapshots
171
Snapshots
Snapshots overview
A OneFS snapshot is a logical pointer to data that is stored on a cluster at a specific point
in time.
A snapshot references a directory on a cluster, including all data stored in the directory
and its subdirectories. If the data referenced by a snapshot is modified, the snapshot
stores a physical copy of the data that was modified. Snapshots are created according to
user specifications or are automatically generated by OneFS to facilitate system
operations.
To create and manage snapshots, you must activate a SnapshotIQ license on the cluster.
Some applications must generate snapshots to function but do not require you to
activate a SnapshotIQ license; by default, these snapshots are automatically deleted
when OneFS no longer needs them. However, if you activate a SnapshotIQ license, you
can retain these snapshots. You can view snapshots generated by other modules without
activating a SnapshotIQ license.
You can identify and locate snapshots by name or ID. A snapshot name is specified by a
user and assigned to the virtual directory that contains the snapshot. A snapshot ID is a
numerical identifier that OneFS automatically assigns to a snapshot.
172
Snapshots
To reduce disk-space usage, snapshots that reference the same directory reference each
other, with older snapshots referencing newer snapshots. If a file is deleted, and several
snapshots reference the file, a single snapshot stores a copy the file, and the other
snapshots reference the file from the snapshot that stored the copy. The reported size of
a snapshot reflects only the amount of data stored by the snapshot and does not include
the amount of data referenced by the snapshot.
Because snapshots do not consume a set amount of storage space, there is no availablespace requirement for creating a snapshot. The size of a snapshot grows according to
how the data referenced by the snapshot is modified. A cluster cannot contain more than
20,000 snapshots.
Snapshot schedules
You can automatically generate snapshots according to a snapshot schedule.
With snapshot schedules, you can periodically generate snapshots of a directory without
having to manually create a snapshot every time. You can also assign an expiration
period that determines when SnapshotIQ deletes each automatically generated
snapshot.
Snapshot aliases
A snapshot alias is an optional, alternative name for a snapshot. If a snapshot is
assigned an alias, and that alias is later assigned to another snapshot, OneFS
automatically removes the alias from the old snapshot and then assigns the alias to the
new snapshot.
Snapshot aliases are most useful when you specify them in a snapshot schedule. When
you specify an alias in a snapshot schedule, SnapshotIQ assigns the alias to the most
recently generated snapshot, enabling you to quickly identify the most recent snapshot
generated according to a schedule.
OneFS uses snapshot aliases internally to identify the most recent snapshot generated by
OneFS operations.
Snapshot schedules
173
Snapshots
It is recommended that you do not disable the snapshot delete job. Disabling the
snapshot delete job prevents unused disk space from being freed and can also cause
performance degradation.
174
Snapshots
Deletion
type
Snapshot
frequency
Ordered
Every hour
deletion
(for mostly
static
data)
Unordered Every other hour
deletion
(for
frequently
Every day
modified
data)
Every week
Every month
Snapshot time
-
Beginning at 12:00
AM Ending at 11:59
AM
Snapshot
expiration
1 month
Beginning at 12:00
AM Ending at 11:59
PM
1 day
At 12:00 AM
1 week
Saturday at 12:00
AM
1 month
3 months
Max snapshots
retained
720
27
File clones
SnapshotIQ enables you to create file clones that share blocks with existing files in order
to save space on the cluster. A file clone usually consumes less space and takes less
time to create than a file copy. Although you can clone files from snapshots, clones are
primarily used internally by OneFS.
The blocks that are shared between a clone and cloned file are contained in a hidden file
called a shadow store. Immediately after a clone is created, all data originally contained
in the cloned file is transferred to a shadow store. Because both files reference all blocks
from the shadow store, the two files consume no more space than the original file; the
clone does not take up any additional space on the cluster. However, if the cloned file or
clone is modified, the file and clone will share only blocks that are common to both of
them, and the modified, unshared blocks will occupy additional space on the cluster.
Over time, the shared blocks contained in the shadow store might become useless if
neither the file nor clone references the blocks. The cluster routinely deletes blocks that
are no longer needed. You can force the cluster to delete unused blocks at any time by
running the shadow store delete job.
Clones cannot contain alternate data streams (ADS). If you clone a file that contains
alternate data streams, the clone will not contain the alternate data streams.
File clones
175
Snapshots
When files that reference shadow stores are replicated to another Isilon cluster or
backed up to a Network Data Management Protocol (NDMP) backup device, the
shadow stores are not transferred to the target Isilon cluster or backup device. The
files are transferred as if they contained the data that they reference from shadow
stores. On the target Isilon cluster or backup device, the files consume the same
amount of space as if they had not referenced shadow stores.
When OneFS creates a shadow store, OneFS assigns the shadow store to a storage
pool of a file that references the shadow store. If you delete the storage pool that a
shadow store resides on, the shadow store is moved to a pool occupied by another
file that references the shadow store.
OneFS does not delete a shadow store block immediately after the last reference to
the block is deleted. Instead, OneFS waits until the ShadowStoreDelete job is run to
delete the unreferenced block. If a large number of unreferenced blocks exist on the
cluster, OneFS might report a negative deduplication savings until the
ShadowStoreDelete job is run.
Shadow stores are protected at least as much as the most protected file that
references it. For example, if one file that references a shadow store resides in a
storage pool with +2 protection and another file that references the shadow store
resides in a storage pool with +3 protection, the shadow store is protected at +3.
Quotas account for files that reference shadow stores as if the files contained the
data referenced from shadow stores; from the perspective of a quota, shadow store
references do not exist. However, if a quota includes data protection overhead, the
quota does not account for the data protection overhead of shadow stores.
Snapshot locks
A snapshot lock prevents a snapshot from being deleted. If a snapshot has one or more
locks applied to it, the snapshot cannot be deleted and is referred to as a locked
snapshot. If the duration period of a locked snapshot expires, OneFS will not delete the
snapshot until all locks on the snapshot have been deleted.
OneFS applies snapshot locks to ensure that snapshots generated by OneFS applications
are not deleted prematurely. For this reason, it is recommended that you do not delete
snapshot locks or modify the duration period of snapshot locks.
A limited number of locks can be applied to a snapshot at a time. If you create snapshot
locks, the limit for a snapshot might be reached, and OneFS could be unable to apply a
snapshot lock when necessary. For this reason, it is recommended that you do not create
snapshot locks.
176
Snapshots
Snapshot reserve
The snapshot reserve enables you to set aside a minimum percentage of the cluster
storage capacity specifically for snapshots. If specified, all other OneFS operations are
unable to access the percentage of cluster capacity that is reserved for snapshots.
Note
The snapshot reserve does not limit the amount of space that snapshots can consume on
the cluster. Snapshots can consume a greater percentage of storage capacity specified by
the snapshot reserve. It is recommended that you do not specify a snapshot reserve.
Inactive
No
Active
Yes
Configure SnapshotIQ
settings
No
Yes
Yes
Yes
Delete snapshots
Yes
Yes
Yes
Yes
View snapshots
Yes
Yes
If you a SnapshotIQ license becomes inactive, you will no longer be able to create new
snapshots, all snapshot schedules will be disabled, and you will not be able to modify
snapshots or snapshot settings. However, you will still be able to delete snapshots and
access data contained in snapshots.
177
Snapshots
directories while the directories are empty. Creating a domain for a directory that contains
less data takes less time.
5. In the Directory Path field, specify the directory that you want to be contained in
snapshots that are generated according to this schedule.
6. Specify how often you want to generate snapshots according to the schedule.
178
Options
Description
Snapshots
Options
Description
Note
A snapshot schedule cannot span multiple days. For example, you cannot specify to
begin generating snapshots at 5:00 PM Monday and end at 5:00 AM Tuesday. To
continuously generate snapshots for a period greater than a day, you must create two
snapshot schedules. For example, to generate snapshots from 5:00 PM Monday to
5:00 AM Tuesday, create one schedule that generates snapshots from 5:00 PM to
11:59 PM on Monday, and another schedule that generates snapshots from 12:00 AM
to 5:00 AM on Tuesday.
7. Optional: To assign an alternative name to the most recent snapshot generated by the
schedule, specify a snapshot alias.
a. Next to Create an Alias, click Yes.
b. To modify the default snapshot alias name, in the Alias Name field, type an
alternative name for the snapshot.
8. Optional: To specify a length of time that snapshots generated according to the
schedule exist on the cluster before they are automatically deleted by OneFS, specify
an expiration period.
a. Next to Snapshot Expiration, click Snapshots expire.
b. Next to Snapshots expire, specify how long you want to retain the snapshots
generated according to the schedule.
9. Click Create.
Create a snapshot
You can create a snapshot of a directory.
Procedure
1. Click Data Protection > SnapshotIQ > Summary.
2. Click Capture a new snapshot.
3. Optional: In the Capture a Snapshot area, in the Snapshot Name field, type a name.
4. In the Directory Path field, specify the directory that you want the snapshot to contain.
5. Optional: To create an alternative name for the snapshot, specify a snapshot alias.
a. Next to Create an Alias, click Yes.
Create a snapshot
179
Snapshots
b. To modify the default snapshot alias name, in the Alias Name field, type an
alternative name for the snapshot.
6. Optional: To assign a time that OneFS will automatically delete the snapshot, specify
an expiration period.
a. Next to Snapshot Expiration, click Snapshot Expires on.
b. In the calendar, specify the day that you want the snapshot to be automatically
deleted.
7. Click Capture.
%A
Description
-
%a
The abbreviated day of the week. For example, if the snapshot is generated
on a Sunday, %a is replaced with Sun.
%B
%b
%C
The first two digits of the year. For example, if the snapshot is created in
2012, %C is replaced with 20.
%c
%d
%e
%F
%G
%g
180
Snapshots
Variable
-
%H
Description
-
The hour. The hour is represented on the 24-hour clock. Single-digit hours are
preceded by a zero. For example, if a snapshot is created at 1:45 AM, %H is
replaced with 01.
%h
%I
The hour represented on the 12-hour clock. Single-digit hours are preceded
by a zero. For example, if a snapshot is created at 1:45 AM, %I is replaced
with 01.
%j
%k
The hour represented on the 24-hour clock. Single-digit hours are preceded
by a blank space.
%l
The hour represented on the 12-hour clock. Single-digit hours are preceded
by a blank space. For example, if a snapshot is created at 1:45 AM, %I is
replaced with 1.
%M
%m
%p
AM or PM.
%{PolicyName} The name of the replication policy that the snapshot was created for. This
variable is valid only if you are specifying a snapshot naming pattern for a
replication policy.
%R
%r
%S
%s
%{SrcCluster} The name of the source cluster of the replication policy that the snapshot was
created for. This variable is valid only if you are specifying a snapshot naming
pattern for a replication policy.
%T
%U
%u
The numerical day of the week. Numbers range from 1 to 7. The first day of
the week is calculated as Monday. For example, if a snapshot is created on
Sunday, %u is replaced with 7.
%V
The two-digit numerical week of the year that the snapshot was created in.
Numbers range from 01 to 53. The first day of the week is calculated as
Monday. If the week of January 1 is four or more days in length, then that
week is counted as the first week of the year.
181
Snapshots
Variable
-
%v
%W
Description
-
The day that the snapshot was created. This variable is equivalent to
specifying %e-%b-%Y.
The two-digit numerical week of the year that the snapshot was created in.
Numbers range from 00 to 53. The first day of the week is calculated as
Monday.
%w
The numerical day of the week that the snapshot was created on. Numbers
range from 0 to 6. The first day of the week is calculated as Sunday. For
example, if the snapshot was created on Sunday, %w is replaced with 0.
%X
The time that the snapshot was created. This variable is equivalent to
specifying %H:%M:%S.
%Y
%y
The last two digits of the year that the snapshot was created in. For example,
if the snapshot was created in 2012, %y is replaced with 12.
%Z
%z
The offset from coordinated universal time (UTC) of the time zone that the
snapshot was created in. If preceded by a plus sign, the time zone is east of
UTC. If preceded by a minus sign, the time zone is west of UTC.
%+
The time and date that the snapshot was created. This variable is equivalent
to specifying %a %b %e %X %Z %Y.
%%
Managing snapshots
You can delete and view snapshots. You can also modify the name, duration period, and
alias of an existing snapshot. However, you cannot modify the data contained in a
snapshot; the data contained in a snapshot is read-only.
Snapshots
If multiple snapshots contain the same directories, deleting older snapshots is more
likely to free disk-space than deleting newer snapshots.
Snapshots that are assigned expiration dates are automatically marked for deletion by
the snapshot daemon. If the daemon is disabled, snapshots will not be automatically
deleted by the system. It is recommended that you do not disable the snapshot daemon.
Delete snapshots
You can delete a snapshot if you no longer want to access the data contained in the
snapshot.
OneFS frees disk space occupied by deleted snapshots when the snapshot delete job is
run. Also, if you delete a snapshot that contains clones or cloned files, data in a shadow
store might no longer be referenced by files on the cluster; OneFS deletes unreferenced
data in a shadow store when the shadow store delete job is run. OneFS routinely runs
both the shadow store delete and snapshot delete jobs. However, you can also manually
run the jobs at any time.
Procedure
1. Click Data Protection > SnapshotIQ > Snapshots.
2. Specify the snapshots that you want to delete.
a. For each snapshot you want to delete, in the Saved File System Snapshots table,
in the row of a snapshot, select the check box.
b. From the Select an action list, select Delete.
c. In the confirmation dialog box, click Delete.
3. Optional: To increase the speed at which deleted snapshot data is freed on the
cluster, run the snapshot delete job.
a. Click Cluster Management > Operations.
b. In the Running Jobs area, click Start Job.
c. From the Job list, select SnapshotDelete.
d. Click Start.
4. Optional: To increase the speed at which deleted data shared between deduplicated
and cloned files is freed on the cluster, run the shadow store delete job.
Run the shadow store delete job only after you run the snapshot delete job.
a. Click Cluster Management > Operations.
b. In the Running Jobs area, click Start Job.
c. From the Job list, select ShadowStoreDelete.
d. Click Start.
Delete snapshots
183
Snapshots
View snapshots
You can view all snapshots.
Procedure
1. Click Data Protection > SnapshotIQ > Snapshots.
2. In the Saved File System Snapshots table, view snapshots.
Snapshot information
You can view information about snapshots, including the total amount of space
consumed by all snapshots.
The following information is displayed in the Saved Snapshots area:
SnapshotIQ Status
Indicates whether a SnapshotIQ license has been activated on the cluster.
Total Number of Saved Snapshots
Indicates the total number of snapshots that exist on the cluster.
Total Number of Snapshots Pending Deletion
Indicates the total number of snapshots that were deleted on the cluster since the
last snapshot delete job was run. The space consumed by the deleted snapshots is
not freed until the snapshot delete job is run again.
Total Number of Snapshot Aliases
Indicates the total number of snapshot aliases that exist on the cluster.
Capacity Used by Saved Snapshots
Indicates the total amount of space consumed by all snapshots.
Snapshots
You can disable and enable access to the snapshots directory for any of these methods
through snapshots settings.
Revert a snapshot
You can revert a directory back to the state it was in when a snapshot was taken.
Before you begin
u
Procedure
1. Click Cluster Management > Operations > Operations Summary.
2. In the Running Jobs area, click Start job.
3. From the Job list, select SnapRevert.
4. Optional: To specify a priority for the job, from the Priority list, select a priority.
Lower values indicate a higher priority. If you do not specify a priority, the job is
assigned the default snapshot revert priority.
5. Optional: To specify the amount of cluster resources the job is allowed to consume,
from the Impact policy list, select an impact policy.
If you do not specify a policy, the job is assigned the default snapshot revert policy.
6. In the Snapshot field, type the name or ID of the snapshot that you want to revert, and
then click Start.
You can access up to 64 snapshots of a directory through Windows explorer, starting with
the most recent snapshot. To access more than 64 snapshots for a directory, access the
cluster through a UNIX command line.
Procedure
1. In Windows Explorer, navigate to the directory that you want to restore or the directory
that contains the file that you want to restore.
2. Right-click the folder, and then click Properties.
3. In the Properties window, click the Previous Versions tab.
4. Select the version of the folder that you want to restore or the version of the folder
that contains the version of the file that you want to restore.
5. Restore the version of the file or directory.
l
To copy the selected directory to another location, click Copy and then specify a
location to copy the directory to.
To restore a specific file, click Open, and then copy the file into the original
directory, replacing the existing copy with the snapshot version.
Revert a snapshot
185
Snapshots
3. Clone a file from the snapshot by running the cp command with the -c option.
For example, the following command clones test.txt from Snapshot2012Jun04:
cp -c /ifs/.snapshot/Snapshot2012Jun04/archive/test.txt \
/ifs/archive/test_clone.text
Snapshots
2. In the Snapshot Schedules table, in the row of the snapshot schedule you want to
modify, click View details.
3. In the Snapshot Schedule Details area, modify snapshot schedule attributes.
4. Next to each snapshot schedule attribute that you modified, click Save.
It is recommended that you do not create, delete, or modify snapshots locks unless you
are instructed to do so by Isilon Technical Support.
Deleting a snapshot lock that was created by OneFS might result in data loss. If you
delete a snapshot lock that was created by OneFS, it is possible that the corresponding
snapshot might be deleted while it is still in use by OneFS. If OneFS cannot access a
snapshot that is necessary for an operation, the operation will malfunction and data loss
might result. Modifying the expiration date of a snapshot lock created by OneFS can also
result in data loss because the corresponding snapshot can be deleted prematurely.
187
Snapshots
snapshot from being automatically deleted, it is recommended that you extend the
duration period of the snapshot by modifying the snapshot.
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
2. Create a snapshot lock by running the isi snapshot locks create command.
For example, the following command applies a snapshot lock to
"SnapshotApril2012", sets the lock to expire in one month, and adds a description of
"Maintenance Lock":
isi snapshot locks create SnapshotApril2012 --expires 1M \
--comment "Maintenance Lock"
It is recommended that you do not modify the expiration dates of snapshot locks.
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
2. Modify a snapshot lock by running the isi snapshot locks modify command.
For example, the following command sets a snapshot lock that is applied to
"SnapshotApril2012" and has an ID of 1 to expire in two days:
isi snapshot locks modify Snapshot2012Apr16 1 --expires 2D
The system prompts you to confirm that you want to delete the snapshot lock.
3. Type yes and then press ENTER.
188
Snapshots
SnapshotIQ settings
SnapshotIQ settings determine how snapshots behave and can be accessed.
The following SnapshotIQ settings can be configured:
Snapshot Scheduling
Determines whether snapshots can be generated.
Note
189
Snapshots
Snapshots
For example, the following command sets the snapshot reserve to 20%:
isi snapshot settings modify --reserve 20
191
CHAPTER 10
Deduplication with SmartDedupe
193
Deduplication overview
The SmartDedupe software module enables you to save storage space on your cluster by
reducing redundant data. Deduplication maximizes the efficiency of your cluster by
decreasing the amount of storage required to store multiple files with similar blocks.
SmartDedupe deduplicates data by scanning an Isilon cluster for identical data blocks.
Each block is 8 KB. If SmartDedupe finds duplicate blocks, SmartDedupe moves a single
copy of the blocks to a hidden file called a shadow store. SmartDedupe then deletes the
duplicate blocks from the original files and replaces the blocks with pointers to the
shadow store.
Deduplication is applied at the directory level, targeting all files and directories
underneath one or more root directories. You can first assess a directory for
deduplication and determine the estimated amount of space you can expect to save. You
can then decide whether to deduplicate the directory. After you begin deduplicating a
directory, you can monitor how much space is saved by deduplication in real time.
SmartDedupe does not deduplicate files that are 32 KB and smaller, because doing so
would consume more cluster resources than the storage savings are worth. Each shadow
store can contain up to 255 blocks. Each block in a shadow store can be referenced
32000 times.
Deduplication jobs
Deduplication is performed by maintenance jobs referred to as deduplication jobs. You
can monitor and control deduplication jobs as you would any other maintenance job on
the cluster. Although the overall performance impact of deduplication is minimal, the
deduplication job consumes 256 MB of memory per node.
When a deduplication job is first run on a cluster, SmartDedupe samples blocks from
each file and creates index entries for those blocks. If the index entries of two blocks
match, SmartDedupe scans the blocks adjacent to the matching pair and then
deduplicates all duplicate blocks. After a deduplication job samples a file once, new
deduplication jobs will not sample the file again until the file is modified.
The first deduplication job you run might take significantly longer to complete than
subsequent deduplication jobs. The first deduplication job must scan all files under the
specified directories to generate the initial index. If subsequent deduplication jobs take a
long time to complete, this most likely indicates that a large amount of data is being
deduplicated. However, it can also indicate that clients are creating a large amount of
new data on the cluster. If a deduplication job is interrupted during the deduplication
process, the job will automatically restart the scanning process from where the job was
interrupted.
It is recommended that you run deduplication jobs when clients are not modifying data
on the cluster. If clients are continually modifying files on the cluster, the amount of
space saved by deduplication is minimal because the deduplicated blocks are constantly
removed from the shadow store. For most clusters, it is recommended that you start a
deduplication job every ten days.
The permissions required to modify deduplication settings are not the same as those
needed to run a deduplication job. Although a user must have the maintenance job
permission to run a deduplication job, the user must have the deduplication permission
to modify deduplication settings. By default, the deduplication job is configured to run at
a low priority.
194
Deduplication considerations
Deduplication can significantly increase the efficiency at which you store data. However,
the effect of deduplication varies depending on the cluster.
You can reduce redundancy on a cluster by running SmartDedupe. Deduplication creates
links that can impact the speed at which you can read from and write to files. In
particular, sequentially reading chunks smaller than 512 KB of a deduplicated file can be
significantly slower than reading the same small, sequential chunks of a nondeduplicated file. This performance degradation applies only if you are reading noncached data. For cached data, the performance for deduplicated files is potentially better
than non-deduplicated files. If you stream chunks larger than 512 KB, deduplication does
not significantly impact the read performance of the file. If you intend on streaming 8 KB
or less of each file at a time, and you do not plan on concurrently streaming the files, it is
recommended that you do not deduplicate the files.
Data replication and backup with deduplication
195
Deduplication is most effective when applied to static or archived files and directories.
The less files are modified, the less negative impact deduplication has on the cluster. For
example, virtual machines often contain several copies of identical files that are rarely
modified. Deduplicating a large number of virtual machines can greatly decrease the
amount of storage space consumed.
SmartDedupe will not deduplicate redundant information within a file. If a file contains
multiple identical data blocks, SmartDedupe will not deduplicate that data unless
another file contains the identical block.
SmartDedupe deduplicates directories that contain iSCSI LUNs the same as other
directories.
When files that reference shadow stores are replicated to another Isilon cluster or
backed up to a Network Data Management Protocol (NDMP) backup device, the
shadow stores are not transferred to the target Isilon cluster or backup device. The
files are transferred as if they contained the data that they reference from shadow
stores. On the target Isilon cluster or backup device, the files consume the same
amount of space as if they had not referenced shadow stores.
When OneFS creates a shadow store, OneFS assigns the shadow store to a storage
pool of a file that references the shadow store. If you delete the storage pool that a
shadow store resides on, the shadow store is moved to a pool occupied by another
file that references the shadow store.
OneFS does not delete a shadow store block immediately after the last reference to
the block is deleted. Instead, OneFS waits until the ShadowStoreDelete job is run to
delete the unreferenced block. If a large number of unreferenced blocks exist on the
cluster, OneFS might report a negative deduplication savings until the
ShadowStoreDelete job is run.
Shadow stores are protected at least as much as the most protected file that
references it. For example, if one file that references a shadow store resides in a
storage pool with +2 protection and another file that references the shadow store
resides in a storage pool with +3 protection, the shadow store is protected at +3.
Quotas account for files that reference shadow stores as if the files contained the
data referenced from shadow stores; from the perspective of a quota, shadow store
references do not exist. However, if a quota includes data protection overhead, the
quota does not account for the data protection overhead of shadow stores.
while the license is inactive. However, you will not be able to deduplicate additional data
until you re-activate the SmartDedupe license.
Managing deduplication
You can manage deduplication on a cluster by first assessing how much space you can
save by deduplicating individual directories. After you determine which directories are
worth deduplicating, you can configure SmartDedupe to deduplicate those directories
specifically. You can then monitor the actual amount of disk space you are saving.
197
198
Deduplication information
You can view the amount of disk space saved by deduplication in the Deduplication
Savings area:
Space Savings
The total amount of physical disk space saved by deduplication, including
protection overhead and metadata. For example, if you have three identical files that
are all 5 GB, the estimated physical saving would be greater than 10 GB, because
deduplication saved space that would have been occupied by file metadata and
protection overhead.
Deduplicated data
The amount of space on the cluster occupied by directories that were deduplicated.
Other data
The amount of space on the cluster occupied by directories that were not
deduplicated.
Deduplication information
199
CHAPTER 11
Data replication with SyncIQ
201
To prevent permissions errors, make sure that ACL policy settings are the same across
source and target clusters.
You can create two types of replication policies: synchronization policies and copy
policies. A synchronization policy maintains an exact replica of the source directory on
the target cluster. If a file or sub-directory is deleted from the source directory, the file or
directory is deleted from the target cluster when the policy is run again.
You can use synchronization policies to fail over and fail back data between source and
target clusters. When a source cluster becomes unavailable, you can fail over data on a
target cluster and make the data available to clients. When the source cluster becomes
available again, you can fail back the data to the source cluster.
A copy policy maintains recent versions of the files that are stored on the source cluster.
However, files that are deleted on the source cluster are not deleted from the target
cluster. Failback is not supported for copy policies. Copy policies are most commonly
used for archival purposes.
Copy policies enable you to remove files from the source cluster without losing those files
on the target cluster. Deleting files on the source cluster improves performance on the
source cluster while maintaining the deleted files on the target cluster. This can be useful
202
if, for example, your source cluster is being used for production purposes and your target
cluster is being used only for archiving.
After creating a job for a replication policy, SyncIQ must wait until the job completes
before it can create another job for the policy. Any number of replication jobs can exist on
a cluster at a given time; however, only five replication jobs can run on a source cluster at
the same time. If more than five replication jobs exist on a cluster, the first five jobs run
while the others are queued to run. The number of replication jobs that a single target
cluster can support concurrently is dependent on the number of workers available on the
target cluster.
You can replicate any number of files and directories with a single replication job. You
can prevent a large replication job from overwhelming the system by limiting the amount
of cluster resources and network bandwidth that data synchronization is allowed to
consume. Because each node in a cluster is able to send and receive data, the speed at
which data is replicated increases for larger clusters.
You can accurately predict when modifications will be made to the data
Configuring a policy to start when changes are made to the source directory can be useful
under the following conditions:
u
For policies that are configured to start whenever changes are made to the source
directory, SyncIQ checks the source directories every ten seconds. SyncIQ does not
account for excluded files or directories when detecting changes, so policies that exclude
files or directories from replication might be run unnecessarily. For example, assume that
newPolicy replicates /ifs/data/media but excludes /ifs/data/media/temp. If a
modification is made to /ifs/data/media/temp/file.txt, SyncIQ will run
newPolicy, but will not replicate /ifs/data/media/temp/file.txt.
If a policy is configured to start whenever changes are made to its source directory, and a
replication job fails, SyncIQ will wait one minute before attempting to run the policy
again. SyncIQ will increase this delay exponentially for each failure up to a maximum
delay of eight hours. You can override the delay by running the policy manually at any
time. After a job for the policy completes successfully, SyncIQ will resume checking the
source directory every ten seconds.
203
target cluster, the mark persists on the target cluster. When a replication policy is run,
SyncIQ checks the mark to ensure that data is being replicated to the correct location.
On the target cluster, you can manually break an association between a replication policy
and target directory. Breaking the association between a source and target cluster causes
the mark on the target cluster to be deleted. You might want to manually break a target
association if an association is obsolete. If you break the association of a policy, the
policy is disabled on the source cluster and you cannot run the policy. If you want to run
the disabled policy again, you must reset the replication policy.
Note
Breaking a policy association causes either a full or differential replication to occur the
next time you run the replication policy. During a full or differential replication, SyncIQ
creates a new association between the source and target clusters. Depending on the
amount of data being replicated, a full or differential replication can take a very long time
to complete.
number of workers per node to increase the speed at which data is replicated to the
target cluster.
You can also reduce resource consumption through file-operation rules that limit the rate
at which replication policies are allowed to send files. However, it is recommended that
you only create file-operation rules if the files you intend to replicate are predictably
similar in size and not especially large.
Replication reports
After a replication job completes, SyncIQ generates a report that contains detailed
information about the job, including how long the job ran, how much data was
transferred, and what errors occurred.
If a replication report is interrupted, SyncIQ might create a subreport about the progress
of the job so far. If the job is then restarted, SyncIQ creates another subreport about the
progress of the job until the job either completes or is interrupted again. SyncIQ creates a
subreport each time the job is interrupted until the job completes successfully. If multiple
subreports are created for a job, SyncIQ combines the information from the subreports
into a single report.
SyncIQ routinely deletes replication reports. You can specify the maximum number of
replication reports that SyncIQ retains and the length of time that SyncIQ retains
replication reports. If the maximum number of replication reports is exceeded on a
cluster, SyncIQ deletes the oldest report each time a new report is created.
You cannot customize the content of a replication report.
Note
If you delete a replication policy, SyncIQ automatically deletes any reports that were
generated for that policy.
Replication snapshots
SyncIQ generates snapshots to facilitate replication, failover, and failback between Isilon
clusters. Snapshots generated by SyncIQ can also be used for archival purposes on the
target cluster.
Replication reports
205
SyncIQ generates source snapshots to ensure that replication jobs do not transfer
unmodified data. When a job is created for a replication policy, SyncIQ checks whether it
is the first job created for the policy. If it is not the first job created for the policy, SyncIQ
compares the snapshot generated for the earlier job with the snapshot generated for the
new job.
SyncIQ replicates only data that has changed since the last time a snapshot was
generated for the replication policy. When a replication job is completed, SyncIQ deletes
the previous source-cluster snapshot and retains the most recent snapshot until the next
job is run.
206
Note
Data failover
Data failover is the process of preparing data on a secondary cluster to be modified by
clients. After you fail over to a secondary cluster, you can redirect clients to modify their
data on the secondary cluster.
Before failover is performed, you must create and run a replication policy on the primary
cluster. You initiate the failover process on the secondary cluster. Failover is performed
per replication policy; to migrate data that is spread across multiple replication policies,
you must initiate failover for each replication policy.
You can use any replication policy to fail over. However, if the action of the replication
policy is set to copy, any file that was deleted on the primary cluster will be present on
the secondary cluster. When the client connects to the secondary cluster, all files that
were deleted on the primary cluster will be available to the client.
If you initiate failover for a replication policy while an associated replication job is
running, the failover operation completes but the replication job fails. Because data
might be in an inconsistent state, SyncIQ uses the snapshot generated by the last
successful replication job to revert data on the secondary cluster to the last recovery
point.
If a disaster occurs on the primary cluster, any modifications to data that were made after
the last successful replication job started are not reflected on the secondary cluster.
When a client connects to the secondary cluster, their data appears as it was when the
last successful replication job was started.
Data failback
Data failback is the process of restoring clusters to the roles they occupied before a
failover operation. After data failback is complete, the primary cluster hosts clients and
replicates data to the secondary cluster for backup.
The first step in the failback process is updating the primary cluster with all of the
modifications that were made to the data on the secondary cluster. The next step in the
failback process is preparing the primary cluster to be accessed by clients. The final step
in the failback process is resuming data replication from the primary to the secondary
cluster. At the end of the failback process, you can redirect users to resume accessing
their data on the primary cluster.
You can fail back data with any replication policy that meets all of the following criteria:
u
The policy does not exclude any files or directories from replication.
207
the last completed replication job started. The RPO is never greater than the time it takes
for two consecutive replication jobs to run and complete.
If a disaster occurs while a replication job is running, the data on the secondary cluster is
reverted to the state it was in when the last replication job completed. For example,
consider an environment in which a replication policy is scheduled to run every three
hours, and replication jobs take two hours to complete. If a disaster occurs an hour after
a replication job begins, the RPO is four hours, because it has been four hours since a
completed job began replicating data.
RTO is the maximum amount of time required to make backup data available to clients
after a disaster. The RTO is always less than or approximately equal to the RPO,
depending on the rate at which replication jobs are created for a given policy.
If replication jobs run continuously, meaning that another replication job is created for
the policy before the previous replication job completes, the RTO is approximately equal
to the RPO. When the secondary cluster is failed over, the data on the cluster is reset to
the state it was in when the last job completed; resetting the data takes an amount of
time proportional to the time it took users to modify the data.
If replication jobs run on an interval, meaning that there is a period of time after a
replication job completes before the next replication job for the policy starts, the
relationship between RTO and RPO depends on whether a replication job is running when
the disaster occurs. If a job is in progress when a disaster occurs, the RTO is roughly
equal to the RPO. However, if a job is not running when a disaster occurs, the RTO is
negligible because the secondary cluster was not modified since the last replication job
ran, and the failover process is almost instantaneous.
208
directories under the included directory are replicated to the target cluster; any
directories that are not contained in an included directory are excluded.
If you both include and exclude directories, any excluded directories must be contained
in one of the included directories; otherwise, the excluded-directory setting has no effect.
For example, consider a policy with the following settings:
u
In this example, the setting that excludes the /ifs/data/archive directory has no
effect because the /ifs/data/archive directory is not under either of the included
directories. The /ifs/data/archive directory is not replicated regardless of whether
the directory is explicitly excluded. However, the setting that excludes the /ifs/data/
media/music/working directory does have an effect, because the directory would be
replicated if the setting was not specified.
In addition, if you exclude a directory that contains the source directory, the excludedirectory setting has no effect. For example, if the root directory of a policy is /ifs/
data, explicitly excluding the /ifs directory does not prevent /ifs/data from being
replicated.
Any directories that you explicitly include or exclude must be contained in or under the
specified root directory. For example, consider a policy in which the specified root
directory is /ifs/data. In this example, you could include both the /ifs/data/
media and the /ifs/data/users/ directories because they are under /ifs/data.
Excluding directories from a synchronization policy does not cause the directories to be
deleted on the target cluster. For example, consider a replication policy that
synchronizes /ifs/data on the source cluster to /ifs/data on the target cluster. If
the policy excludes /ifs/data/media from replication, and /ifs/data/media/
file exists on the target cluster, running the policy does not cause /ifs/data/
media/file to be deleted from the target cluster.
209
210
File name
Includes or excludes files based on the file name. You can specify to include or
exclude full or partial names that contain specific text.
The following wildcard characters are accepted:
Note
Alternatively, you can filter file names by using POSIX regular-expression (regex) text.
Isilon clusters support IEEE Std 1003.2 (POSIX.2) regular expressions. For more
information about POSIX regular expressions, see the BSD man pages.
Table 14 Replication file matching wildcards
Wildcard Description
*
[ ]
Path
Includes or excludes files based on the file path. This option is available for copy
policies only.
You can specify to include or exclude full or partial paths that contain specified text.
You can also include the wildcard characters *, ?, and [ ].
Size
Includes or excludes files based on their size.
Note
211
Type
Includes or excludes files based on one of the following file-system object types:
u
Soft link
Regular file
Directory
Click Connect to only the nodes in the subnet and pool if the target cluster name
specifies a SmartConnect zone.
3. Specify which nodes you want replication policies to connect to when a policy is run.
Options
Description
Note
SyncIQ does not support dynamically allocated IP address pools. If a replication job
connects to a dynamically allocated IP address, SmartConnect might reassign the
address while a replication job is running, which would disconnect the job and cause
it to fail.
4. Click Submit.
212
Source directory
File-criteria statement
Target directory
To copy all files from the source directory to the target directory, click Copy.
Note
To copy all files from the source directory to the target directory and delete any
files on the target directory that are not in the source directory, click Synchronize.
6. In the Run job area, specify whether replication jobs will be run.
Options
Description
a. Click On a schedule.
b. Specify a schedule.
If you configure a replication policy to run more
than once a day, you cannot configure the interval
to span across two calendar days. For example,
you cannot configure a replication policy to run
every hour starting at 7:00 PM and ending at 1:00
AM.
213
To include a directory, in the Included Directories area, click Add a directory path.
To exclude a directory, in the Excluded Directories area, click Add a directory path.
3. Optional: Prevent specific files from being replicated by specifying file matching
criteria.
a. In the File Matching Criteria area, select a filter type.
b. Select an operator.
c. Type a value.
Files that do not meet the specified criteria will not be replicated to the target cluster.
For example, if you specify File Type doesn't match .txt, SyncIQ will not
replicate any files with the .txt file extension. If you specify Created after
08/14/2013, SyncIQ will not replicate any files created before August 14th, 2013.
If you want to specify more than one file matching criterion, you can control how the
criteria relate to each other by clicking either Add an "Or" condition or Add an "And"
condition.
4. Specify which nodes you want the replication policy to connect to when the policy is
run.
Options
Description
Note
SyncIQ does not support dynamically allocated IP address pools. If a replication job
connects to a dynamically allocated IP address, SmartConnect might reassign the
address while a replication job is running, which would disconnect the job and cause
it to fail.
After you finish
The next step in the process of creating a replication policy is specifying the target
directory.
214
The fully qualified domain name of any node in the target cluster.
localhost
SyncIQ does not support dynamically allocated IP address pools. If a replication job
connects to a dynamically allocated IP address, SmartConnect might reassign the
address while a replication job is running, which would disconnect the job and cause
it to fail.
2. In the Target Directory field, type the absolute path of the directory on the target
cluster that you want to replicate data to.
CAUTION
If you specify an existing directory on the target cluster, ensure that the directory is
not the target of another replication policy. If this is a synchronization policy, ensure
that the directory is empty. All files are deleted from the target of a synchronization
policy the first time the policy is run.
If the specified target directory does not already exist on the target cluster, the
directory is created the first time the job is run. It is recommended that you do not
specify the /ifs directory. If you specify the /ifs directory, the entire target cluster
is set to a read-only state, preventing you from storing any other data on the cluster.
If this is a copy policy, and files in the target directory share the same name as files in
the source directory, the target directory files are overwritten when the job is run.
3. If you want replication jobs to connect only to the nodes included in the SmartConnect
zone specified by the target cluster, click Connect only to the nodes within the target
cluster SmartConnect Zone.
After you finish
The next step in the process of creating a replication policy is specifying policy target
snapshot settings.
215
Procedure
1. To create archival snapshots on the target cluster, in the Target Snapshots area, click
Capture snapshots on the target cluster.
2. Optional: To modify the default alias of the last snapshot created according to the
replication policy, in the Snapshot Alias Name field, type a new alias.
You can specify the alias name as a snapshot naming pattern. For example, the
following naming pattern is valid:
%{PolicyName}-on-%{SrcCluster}-latest
3. Optional: To modify the snapshot naming pattern for snapshots created according to
the replication policy, in the Snapshot Naming Pattern field, type a naming pattern.
Each snapshot generated for this replication policy is assigned a name based on this
pattern.
For example, the following naming pattern is valid:
%{PolicyName}-from-%{SrcCluster}-at-%H:%M-on-%m-%d-%Y
Do not modify the default setting without consulting Isilon Technical Support.
2. Optional: From the Log Level list, select the level of logging you want SyncIQ to
perform for replication jobs.
The following log levels are valid, listed from least to most verbose:
l
Click Error.
Click Notice.
3. Optional: If you want SyncIQ to perform a checksum on each file data packet that is
affected by the replication policy, select the Validate File Integrity check box.
If you enable this option, and the checksum values for a file data packet do not
match, SyncIQ retransmits the affected packet.
216
4. Optional: To modify the length of time SyncIQ retains replication reports for the policy,
in the Keep Reports For area, specify a length of time.
After the specified expiration period has passed for a report, SyncIQ automatically
deletes the report.
Some units of time are displayed differently when you view a report than how they
were originally entered. Entering a number of days that is equal to a corresponding
value in weeks, months, or years results in the larger unit of time being displayed. For
example, if you enter a value of 7 days, 1 week appears for that report after it is
created. This change occurs because SyncIQ internally records report retention times
in seconds and then converts them into days, weeks, months, or years.
5. Optional: Specify whether to record information about files that are deleted by
replication jobs by selecting one of the following options:
l
217
You can assess only replication policies that have never been run before.
Procedure
1. Click Data Protection > SyncIQ > Policies.
2. In the SyncIQ Policies table, in the row of a replication policy, from the Actions
column, select Assess Sync.
3. Click Data Protection > SyncIQ > Summary.
4. After the job completes, in the SyncIQ Recent Reports table, in the row of the
replication job, click View Details.
The report displays the total amount of data that would have been transferred in the
Total Data field.
218
219
Elapsed
How much time has elapsed since the job started.
Transferred
The number of files that have been transferred, and the total size of all transferred
files.
Source Directory
The path of the source directory on the source cluster.
Target Host
The target directory on the target cluster.
Actions
Displays any job-related actions that you can perform.
Although you cannot fail over or fail back SmartLock directories, you can recover
SmartLock directories on a target cluster. After you recover SmartLock directories, you can
migrate them back to the source cluster.
is useful if the primary cluster becomes available before data is modified on the
secondary cluster or if you failed over to a secondary cluster for testing purposes.
Before you begin
Fail over a replication policy.
Reverting a failover operation does not migrate modified data back to the primary cluster.
To migrate data that clients have modified on the secondary cluster, you must fail back to
the primary cluster.
Complete the following procedure for each replication policy that you want to fail over:
Procedure
1. Click Data Protection > SyncIQ > Local Targets.
2. In the SyncIQ Local Targets table, in the row for a replication policy, from the Actions
column, select Disallow Writes.
3. On the secondary cluster, replicate data to the primary cluster by using the mirror
policies.
You can replicate data either by manually starting the mirror policies or by modifying
the mirror policies and specifying a schedule.
4. Prevent clients from accessing the secondary cluster and then run each mirror policy
again.
To minimize impact to clients, it is recommended that you wait until client access is
low before preventing client access to the cluster.
5. On the primary cluster, click Data Protection > SyncIQ > Local Targets.
6. In the SyncIQ Local Targets table, from the Actions column, select Allow Writes for
each mirror policy.
7. On the secondary cluster, click Data Protection > SyncIQ > Policies.
8. In the SyncIQ Policies table, from the Actions column, select Resync-prep for each
mirror policy.
After you finish
Redirect clients to begin accessing the primary cluster.
221
If the last replication job completed successfully and a replication job is not
currently running, select Allow Writes.
If a replication job is currently running, wait until the replication job completes,
and then select Allow Writes.
If the primary cluster became unavailable while a replication job was running,
select Break Association.
3. If you clicked Break Association, restore any files that are left in an inconsistent state.
a. Delete all files that are not committed to a WORM state from the target directory.
b. Copy all files from the failover snapshot to the target directory.
Failover snapshots are named according to the following naming pattern:
SIQ-Failover-<policy-name>-<year>-<month>-<day>_<hour>-<minute><second>
222
The source directory is the SmartLock directory that you are migrating.
The target directory is an empty SmartLock directory. The source and target
directories must be of the same SmartLock type. For example, if the target
directory is a compliance directory, the source must also be a compliance
directory.
2. Replicate data to the target cluster by running the policies you created.
You can replicate data either by manually starting the policies or by specifying a policy
schedule.
3. Optional: To ensure that SmartLock protection is enforced for all files, commit all files
in the SmartLock source directory to a WORM state.
Because autocommit information is not transferred to the target cluster, files that
were scheduled to be committed to a WORM state on the source cluster will not be
scheduled to be committed at the same time on the target cluster. To ensure that all
files are retained for the appropriate time period, you can commit all files in target
SmartLock directories to a WORM state.
For example, the following command automatically commits all files in /ifs/data/
smartlock to a WORM state after one minute.
isi smartlock modify --path /ifs/data/smartlock --autocommit 1n
This step is unnecessary if you have not configured an autocommit time period for the
SmartLock directory being replicated.
4. Prevent clients from accessing the source cluster and run the policy that you created.
To minimize impact to clients, it is recommended that you wait until client access is
low before preventing client access to the cluster.
5. On the target cluster, click Data Protection > SyncIQ > Local Targets.
6. In the SyncIQ Local Targets table, in the row of each replication policy, from the
Actions column, select Allow Writes.
7. Optional: If any SmartLock directory configuration settings, such as an autocommit
time period, were specified for the source directories of the replication policies, apply
those settings to the target directories.
8. Optional: Delete the copy of your SmartLock data on the source cluster.
If the SmartLock directories are compliance directories or enterprise directories with
the privileged delete functionality permanently disabled, you cannot recover the
space consumed by the source SmartLock directories until all files are released from a
WORM state. If you want to free the space before files are released from a WORM
state, contact Isilon Technical Support for information about reformatting your cluster.
223
Source directory
File-criteria statement
Target cluster
This applies only if you target a different cluster. If you modify the IP or domain name
of a target cluster, and then modify the replication policy on the source cluster to
match the new IP or domain name, a full replication is not performed.
Target directory
Procedure
1. Click Data Protection > SyncIQ > Policies.
2. In the SyncIQ Policies table, in the row for a policy, click View/Edit.
3. In the View SyncIQ Policy Details dialog box, click Edit Policy.
4. Modify the settings of the replication policy, and then click Save Changes.
If you disable a replication policy while an associated replication job is running, the
running job is not interrupted. However, the policy will not create another job until the
policy is enabled.
224
Procedure
1. Click Data Protection > SyncIQ > Policies.
2. In the SyncIQ Policies table, in the row for a replication policy, select either Enable
Policy or Disable Policy.
If neither Enable Policy nor Disable Policy appears, verify that a replication job is not
running for the policy. If an associated replication job is not running, ensure that the
SyncIQ license is active on the cluster.
225
Action
Determines the how the policy replicates data. All policies copy files from the source
directory to the target directory and update files in the target directory to match files
on the source directory. The action determines how deleting a file on the source
directory affects the target. The following values are valid:
Copy
If a file is deleted in the source directory, the file is not deleted in the target
directory.
Synchronize
Deletes files in the target directory if they are no longer present on the source.
This ensures that an exact replica of the source directory is maintained on the
target cluster.
Run job
Determines whether jobs are run automatically according to a schedule or only when
manually specified by a user.
Last Successful Run
Displays the last time that a replication job for the policy completed successfully.
Last Started
Displays the last time that the policy was run.
Source Root Directory
The full path of the source directory. Data is replicated from the source directory to
the target directory.
Included Directories
Determines which directories are included in replication. If one or more directories
are specified by this setting, any directories that are not specified are not replicated.
Excluded Directories
Determines which directories are excluded from replication. Any directories specified
by this setting are not replicated.
File Matching Criteria
Determines which files are excluded from replication. Any files that do not meet the
specified criteria are not replicated.
Restrict Source Nodes
Determines whether the policy can run on all nodes on the source cluster or run only
on specific nodes.
Target Host
The IP address or fully qualified domain name of the target cluster.
Target Directory
The full path of the target directory. Data is replicated to the target directory from the
source directory.
Restrict Target Nodes
Determines whether the policy can connect to all nodes on the target cluster or can
connect only to specific nodes.
Capture Snapshots
Determines whether archival snapshots are generated on the target cluster.
Snapshot Alias Name
Specifies an alias for the latest archival snapshot taken on the target cluster.
Snapshot Naming Pattern
Specifies how archival snapshots are named on the target cluster.
226
Snapshot Expiration
Specifies how long archival snapshots are retained on the target cluster before they
are automatically deleted by the system.
Workers Threads Per Node
Specifies the number of workers per node that are generated by OneFS to perform
each replication job for the policy.
Log Level
Specifies the amount of information that is recorded for replication jobs.
More verbose options include all information from less verbose options. The
following list describes the log levels from least to most verbose:
Notice
Includes job and process-level activity, including job starts, stops, and worker
coordination information. This is the recommended log level.
Error
Includes events related to specific types of failures.
Network Activity
Includes more job-level activity and work-item information, including specific
paths and snapshot names.
File Activity
Includes a separate event for each action taken on a file. Do not select this
option without first consulting Isilon Technical Support.
Replication logs are typically used for debugging purposes. If necessary, you can log
in to a node through the command-line interface and view the contents of
the /var/log/isi_migrate.log file on the node.
Validate File Integrity
Determines whether OneFS performs a checksum on each file data packet that is
affected by a replication job. If a checksum value does not match, OneFS retransmits
the affected file data packet.
Keep Reports For
Specifies how long replication reports are kept before they are automatically deleted
by OneFS.
Log Deletions on Synchronization
Determines whether OneFS records when a synchronization job deletes files or
directories on the target cluster.
The following replication policy fields are available only through the OneFS command-line
interface.
Source Subnet
Specifies whether replication jobs connect to any nodes in the cluster or if jobs can
connect only to nodes in a specified subnet.
Source Pool
Specifies whether replication jobs connect to any nodes in the cluster or if jobs can
connect only to nodes in a specified pool.
Password Set
Specifies a password to access the target cluster.
Report Max Count
Specifies the maximum number of replication reports that are retained for this
policy.
227
Disabling this option could result in data loss. It is recommended that you consult
Isilon Technical Support before disabling this option.
Resolve
Determines whether you can manually resolve the policy if a replication job
encounters an error.
228
To cancel a specific job, in the row for a replication job, select Cancel Running Job.
To cancel all jobs targeting the local cluster, select the check box to the left of
Policy Name and then select Cancel Selection from the Select a bulk action list.
After a replication policy is reset, SyncIQ performs a full or differential replication the
next time the policy is run. Depending on the amount of data being replicated, a full or
differential replication can take a very long time to complete.
Procedure
1. Click Data Protection > SyncIQ > Local Targets.
2. In the SyncIQ Local Targets table, in the row for a replication policy, select Break
Association.
3. In the Confirm dialog box, click Yes.
229
Actions
Displays any job-related actions that you can perform.
230
231
3. In the Number of Reports to Keep Per Policy field, type the maximum number of
reports you want to retain at a time for a replication policy.
4. Click Submit.
232
Sync Type
The action that was performed by the replication job.
Initial Sync
Indicates that either a differential or a full replication was performed.
Incremental Sync
Indicates that only modified files were transferred to the target cluster.
Failover / Failback Allow Writes
Indicates that writes were enabled on a target directory of a replication policy.
Failover / Failback Disallow Writes
Indicates that an allow writes operation was undone.
Failover / Failback Resync Prep
Indicates that an association between files on the source cluster and files on
the target cluster was created. This is the first step in the failback preparation
process.
Failover / Failback Resync Prep Domain Mark
Indicates that a SyncIQ domain was created for the source directory. This is the
second step in the failback preparation process.
Failover / Failback Resync Prep Restore
Indicates that a source directory was restored to the last recovery point. This is
the third step in the failback preparation process.
Failover / Failback Resync Prep Finalize
Indicates that a mirror policy was created on the target cluster. This is the last
step in the failback preparation process.
Upgrade
Indicates that a policy-conversion replication occurred after upgrading the
OneFS operating system or merging policies.
Source
The path of the source directory on the source cluster.
Target
The IP address or fully qualified domain name of the target cluster.
Actions
Displays any report-related actions that you can perform.
233
you can reset the replication policy. However, resetting the policy causes a full or
differential replication to be performed the next time the policy is run.
Note
Depending on the amount of data being synchronized or copied, a full and differential
replications can take a very long time to complete.
Depending on the amount of data being replicated, a full or differential replication can
take a very long time to complete. Reset a replication policy only if you cannot fix the
issue that caused the replication error. If you fix the issue that caused the error, resolve
the policy instead of resetting the policy.
Procedure
1. Click Data Protection > SyncIQ > Policies.
2. In the SyncIQ Policies table, in the row for a policy, select Reset Sync State.
234
3. Run the policy by running the isi sync jobs start command.
For example, the following command runs newPolicy:
isi sync jobs start newPolicy
235
CHAPTER 12
Data layout with FlexProtect
FlexProtect overview............................................................................................238
File striping......................................................................................................... 238
Requested data protection.................................................................................. 238
FlexProtect data recovery.....................................................................................239
Requesting data protection................................................................................. 240
Requested protection settings.............................................................................240
Requested protection disk space usage.............................................................. 241
237
FlexProtect overview
An Isilon cluster is designed to continuously serve data, even when one or more
components simultaneously fail. OneFS ensures data availability by striping or mirroring
data across the cluster. If a cluster component fails, data stored on the failed component
is available on another component. After a component failure, lost data is restored on
healthy components by the FlexProtect proprietary system.
Data protection is specified at the file level, not the block level, enabling the system to
recover data quickly. Because all data, metadata, and parity information is distributed
across all nodes, the cluster does not require a dedicated parity node or drive. This
ensures that no single node limits the speed of the rebuild process.
File striping
OneFS uses the internal network to automatically allocate and stripe data across nodes
and disks in the cluster. OneFS protects data as the data is being written. No separate
action is necessary to stripe data.
OneFS breaks files into smaller logical chunks called stripes before writing the files to
disk; the size of each file chunk is referred to as the stripe unit size. Each OneFS block is
8 KB, and a stripe unit consists of 16 blocks, for a total of 128 KB per stripe unit. During a
write, OneFS breaks data into stripes and then logically places the data in a stripe unit.
As OneFS stripes data across the cluster, OneFS fills the stripe unit according to the
number of nodes and protection level.
OneFS can continuously reallocate data and make storage space more usable and
efficient. As the cluster size increases, OneFS stores large files more efficiently.
Smartfail
OneFS protects data stored on failing nodes or drives through a process called
smartfailing.
During the smartfail process, OneFS places a device into quarantine. Data stored on
quarantined devices is read only. While a device is quarantined, OneFS reprotects the
data on the device by distributing the data to other devices. After all data migration is
complete, OneFS logically removes the device from the cluster, the cluster logically
changes its width to the new configuration, and the node or drive can be physically
replaced.
OneFS smartfails devices only as a last resort. Although you can manually smartfail
nodes or drives, it is recommended that you first consult Isilon Technical Support.
Occasionally a device might fail before OneFS detects a problem. If a drive fails without
being smartfailed, OneFS automatically starts rebuilding the data to available free space
on the cluster. However, because a node might recover from a failure, if a node fails,
OneFS does not start rebuilding data unless the node is logically removed from the
cluster.
Node failures
Because node loss is often a temporary issue, OneFS does not automatically start
reprotecting data when a node fails or goes offline. If a node reboots, the file system does
not need to be rebuilt because it remains intact during the temporary failure.
If you configure N+1 data protection on a cluster, and one node fails, all of the data is still
accessible from every other node in the cluster. If the node comes back online, the node
rejoins the cluster automatically without requiring a full rebuild.
To ensure that data remains protected, if you physically remove a node from the cluster,
you must also logically remove the node from the cluster. After you logically remove a
node, the node automatically reformats its own drives, and resets itself to the factory
FlexProtect data recovery
239
default settings. The reset occurs only after OneFS has confirmed that all data has been
reprotected. You can logically remove a node using the smartfail process. It is important
that you smartfail nodes only when you want to permanently remove a node from the
cluster.
If you remove a failed node before adding a new node, data stored on the failed node
must be rebuilt in the free space in the cluster. After the new node is added, OneFS
distributes the data to the new node. It is more efficient to add a replacement node to the
cluster before failing the old node because OneFS can immediately use the replacement
node to rebuild the data stored on the failed node.
For 4U Isilon IQ X-Series and NL-Series nodes, and IQ 12000X/EX 12000 combination
platforms, the minimum cluster size of three nodes requires a minimum of N+2:1.
240
N+1
Minimum number of
nodes required
-
Definition
-
N+2:1
N+2
Requested protection
setting
-
Minimum number of
nodes required
-
N+3:1
Definition
-
N+3
N+4
Nx (Data mirroring)
+1
-
+2:1
-
2 +1 (33%) 4 + 2 (33%)
+2
-
+3:1
-
3x
+3
-
+4
-
3 +1 (25%) 6 + 2 (25%)
2 + 2 (50%)
9 + 3 (25%)
4x
4 +1 (20%) 8 + 2 (20%)
3 + 2 (40%)
12 + 3
(20%)
4x
5x
5 +1 (17%) 10 + 2 (17%)
4 + 2 (33%)
15 + 3
(17%)
3 + 3 (50%) 5x
6 +1 (14%) 12 + 2 (14%)
5 + 2 (29%)
15 + 3
(17%)
4 + 3 (43%) 5x
15 + 3
(17%)
5 + 3 (38%) 4 + 4
(50%)
8 +1 (11%) 16 + 2 (11%)
15 + 3
(17%)
6 + 3 (33%) 5 + 4
(44%)
7 + 2 (22%)
241
Number
of nodes
-
+1
-
10
+2:1
+2
9 +1 (10%) 16 + 2 (11%)
8 + 2 (20%)
+3:1
-
15 + 3
(17%)
+3
-
+4
-
7 + 3 (30%) 6 + 4
(40%)
12
11 +1 (8%) 16 + 2 (11%)
10 + 2
(17%)
15 + 3
(17%)
9 + 3 (25%) 8 + 4
(33%)
14
13 + 1 (7%) 16 + 2 (11%)
12 + 2
(14%)
15 + 3
(17%)
11 + 3
(21%)
10 + 4
(29%)
16
15 + 1 (6%) 16 + 2 (11%)
14 + 2
(13%)
15 + 3
(17%)
13 + 3
(19%)
12 + 4
(25%)
18
16 + 1 (6%) 16 + 2 (11%)
16 + 2
(11%)
15 + 3
(17%)
15 + 3
(17%)
14 + 4
(22%)
20
16 + 1 (6%) 16 + 2 (11%)
16 + 2
(11%)
16 + 3
(16%)
16 + 3
(16%)
16 + 4
(20%)
30
16 + 1 (6%) 16 + 2 (11%)
16 + 2
(11%)
16 + 3
(16%)
16 + 3
(16%)
16 + 4
(20%)
The parity overhead for mirrored data protection is not affected by the number of nodes in
the cluster. The following table describes the parity overhead for requested mirrored
protection.
2x
-
242
3x
-
4x
-
5x
-
6x
-
7x
-
8x
-
CHAPTER 13
NDMP backup
NDMP backup
243
NDMP backup
NDMP backup
then connect the Fibre Channel switch to two Fibre Channel ports, OneFS creates two
entries for the device, one for each path.
Note
If you perform an NDMP two-way backup operation, you must assign static IP addresses
to the Backup Accelerator node. If you connect to the cluster through a data management
application (DMA), you must connect to the IP address of a Backup Accelerator node. If
you perform an NDMP three-way backup, you can connect to any node in the cluster.
DMA
Symantec NetBackup
Supported
Yes
EMC Networker
Yes
EMC Avamar
No
Commvault Simpana
No
Yes
Dell NetVault
Yes
ASG-Time Navigator
Yes
245
NDMP backup
In a level 10 NDMP backup, only data changed since the most recent incremental
(level 1-9) backup or the last level 10 backup is copied. By repeating level 10
backups, you can be assured that the latest versions of files in your data set are
backed up without having to run a full backup.
u
Supported DMAs
NDMP backups are coordinated by a data management application (DMA) that runs on a
backup server.
OneFS supports the following DMAs:
246
Symantec NetBackup
EMC NetWorker
EMC Avamar
Dell NetVault
ASG-Time Navigator
NDMP backup
LTO-3
LTO-4
LTO-5
OneFS does not back up file system configuration data, such as file protection level
policies and quotas.
OneFS does not support multiple concurrent backups onto the same tape.
OneFS does not support restoring data from a file system other than OneFS. However,
you can migrate data via the NDMP protocol from a NetApp or EMC VNX storage
system to OneFS.
Backup Accelerator nodes cannot interact with more than 1024 device paths,
including the paths of tape and media changer devices. For example, if each device
has four paths, you can connect 256 devices to a Backup Accelerator node. If each
device has two paths, you can connect 512 devices.
OneFS does not support more than 64 concurrent NDMP sessions per Backup
Accelerator node.
Install the latest patches for OneFS and your data management application (DMA).
If you are backing up multiple directories that contain small files, set up a separate
schedule for each directory.
If you are performing three-way NDMP backups, run multiple NDMP sessions on
multiple nodes in your Isilon cluster.
Restore files through Direct Access Restore (DAR) and Directory DAR (DDAR). This is
especially recommended if you restore files frequently. However, it is recommended
that you do not use DAR to restore a full backup or a large number of files, as DAR is
better suited to restoring smaller numbers of files.
Use the largest tape record size available for your version of OneFS. The largest tape
record size for OneFS versions 6.5.5 and later is 256 KB. The largest tape record size
for versions of OneFS earlier than 6.5.5 is 128 KB.
NDMP hardware support
247
NDMP backup
If possible, do not include or exclude files from backup. Including or excluding files
can affect backup performance, due to filtering overhead.
Limit the number of files in a directory. Distribute files across multiple directories
instead of including a large number of files in a single directory.
Networking recommendations
u
Connect NDMP sessions only through SmartConnect zones that are exclusively used
for NDMP backup.
Configure multiple policies when scheduling backup operations, with each policy
capturing a portion of the file system. Do not attempt to back up the entire file system
through a single policy.
This is recommended only if you are backing up a significant amount of data. Running
four concurrent streams might not be necessary for smaller backups.
u
Attach more Backup Accelerator nodes to larger clusters. The recommended number
of Backup Accelerator nodes is listed in the following table.
Table 16 Nodes per Backup Accelerator node
X-Series
NL-Series
S-Series
Attach more Backup Accelerator nodes if you are backing up to more tape devices.
The following table lists the recommended number of tape devices per backup
accelerator node:
Table 17 Tape devices per Backup Accelerator node
Tape device type Recommended number of tape devices per Backup Accelerator node
-
LTO-5
LTO-4
LTO-3
DMA-specific recommendations
u
248
NDMP backup
Character
-
Description
-
Example
-
archive*
/ifs/data/data_store_a
/ifs/data/data_store_c
user_?
/ifs/data/user_1
/ifs/data/user_2
Includes a blank
space
user\ 1
/ifs/data/user 1
[]
/ifs/data/data_store_8
Unanchored patterns such as home or user1 target a string of text that might belong to
many files or directories. Anchored patterns target specific file pathnames, such as ifs/
data/home. You can include or exclude either type of pattern.
For example, suppose you want to back up the /ifs/data/home directory, which
contains the following files and directories:
u
/ifs/data/home/user1/file.txt
/ifs/data/home/user2/user1/file.txt
/ifs/data/home/user3/other/file.txt
/ifs/data/home/user4/emptydirectory
If you simply include the /ifs/data/home directory, all files and directories, including
emptydirectory would be backed up.
If you specify both include and exclude patterns, any excluded files or directories under
the included directories would not be backed up. If the excluded directories are not found
in any of the included directories, the exclude specification would have no effect.
Note
249
NDMP backup
From the DMA vendor list, select the name of the DMA vendor to manage backup
operations.
If your DMA vendor is not included in the list, select generic. However, note that
any vendors not included on the list are not officially supported and might not
function as expected.
4. Click Add administrator to add an NDMP user account through which your DMA can
access the cluster.
a. In the Add Administrator dialog box, in the Name field, type a name for the
account.
b. In the Password and Confirm password fields, type a password for the account.
c. Click Submit.
250
NDMP backup
251
NDMP backup
252
NDMP backup
Setting
Name
State
Description
A device name assigned by OneFS.
Indicates whether the device is in use. If data is currently being backed up to
or restored from the device, Read/Write appears. If the device is not in
use, Closed appears.
WWN
Product
The name of the device vendor and the model name or number of the device.
Serial Number
Paths
The name of the Backup Accelerator node that the device is attached to and
the port number or numbers to which the device is connected.
LUN
Port ID
The port ID of the device that binds the logical device to the physical device.
253
NDMP backup
Setting
WWPN
Description
The world wide port name (WWPN) of the port on the tape or media changer
device.
254
Setting
Port
Description
The name of the Backup Accelerator node, and the number of the port.
NDMP backup
Setting
Topology
Description
The type of Fibre Channel topology that the port is configured to support..
Options are:
Point to Point
A single backup device or Fibre Channel switch directly connected to
the port.
Loop
Multiple backup devices connected to a single port in a circular
formation.
Auto
Automatically detects the topology of the connected device. This is the
recommended setting, and is required if you are using a switchedfabric topology.
WWNN
The world wide node name (WWNN) of the port. This name is the same for
each port on a given node.
WWPN
The world wide port name (WWPN) of the port. This name is unique to the
port.
Rate
The rate at which data is sent through the port. The rate can be set to 1
Gb/s, 2 Gb/s, 4 Gb/s, 8 Gb/s, and Auto. 8 Gb/s is available for
A100 nodes only. If set to Auto, OneFS automatically negotiates with the
DMA to determine the rate. Auto is the recommended setting.
255
NDMP backup
Item
Session
Description
The unique identification number that OneFS assigned to the
session.
Elapsed
Transferred
Throughput
The average throughput of the session over the past five minutes.
Client/Remote
Mover/Data
The current state of the data mover and the data server. The first
word describes the activity of the data mover. The second word
describes the activity of the data server.
The data mover and data server send data to and receive data from
each other during backup and restore operations. The data mover is
a component of the backup server that receives data during
backups and sends data during restore operations. The data server
is a component of OneFS that sends data during backups and
receives information during restore operations.
The following states might appear:
Active
The data mover or data server is currently sending or receiving
data.
Paused
The data mover is temporarily unable to receive data. While the
data mover is paused, the data server cannot send data to the
data mover. The data server cannot be paused.
Idle
The data mover or data server is not sending or receiving data.
Listen
The data mover or data server is waiting to connect to the data
server or data mover.
Operation
Backup (0-10)
Indicates that data is currently being backed up to a media
server. The number indicates the level of NDMP backup.
256
NDMP backup
Item
Description
Restore
Indicates that data is currently being restored from a media
server.
Source/Destination
Device
Mode
Read/Write
OneFS is reading and writing data during a backup operation.
Read
OneFS is reading data during a restore operation.
Raw
The DMA has access to tape drives, but the drives do not
contain writable tape media.
257
NDMP backup
Procedure
1. Configure the client and the directory path that you want to back up as you would
normally.
2. In the Client Properties dialog box, enable restartable backups.
a. On the General screen, click the Checkpoint enabled checkbox.
b. Specify File in the Checkpoint granularity drop-down list.
3. In the Application information field, type any NDMP variables that you want to specify.
The following specifies a checkpoint interval of 1 GB.
CHECKPOINT_INTERVAL_IN_BYTES=1GB
2. To view detailed information about a specific backup context, run the isi ndmp
extensions contexts view command.
The following command displays detailed information about a backup context with an
ID of 792eeb8a-8784-11e2-aa70-0025904e91a4:
isi ndmp extensions contexts view 792eeb8a-8784-11e2aa70-0025904e91a4
258
NDMP backup
Tested configurations
7.1.0.1
7.0.2.5
6.6.5.26
* The tape drive sharing function is not supported in the OneFS 7.0.1 release.
EMC NetWorker refers to the tape drive sharing capability as DDS (dynamic drive sharing).
Symantec NetBackup uses the term SSO (shared storage option). Consult your DMA
vendor documentation for configuration instructions.
259
NDMP backup
through your data management application (DMA). For more information about specifying
NDMP environment variables through your DMA, see your DMA documentation.
3. Optional: To remove a default NDMP setting for a directory, run the isi ndmp
settings variables delete command:
For example, the following command removes the default file history format
for /ifs/data/media:
isi ndmp settings variables delete /ifs/data/media --name HIST
260
NDMP backup
Environment variable
Valid values
Default
TIMESTAMP
Description
BACKUP_MODE=
TIMESTAMP
SNAPSHOT
FILESYSTEM=
<file-path>
None
LEVEL=
<integer>
0
Performs a full NDMP
backup.
1-9
Performs an incremental
backup at the specified level.
10
Performs unlimited
incremental backups.
UPDATE=
Y
N
Y
OneFS updates the dump
dates file.
N
OneFS does not update the
dump dates file.
HIST=
<file-historyformat>
D
Specifies dir/node file
history.
261
NDMP backup
Environment variable
Valid values
Default
Description
F
Specifies path-based file
history.
Y
Specifies the default file
history format determined by
your NDMP backup settings.
N
Disables file history.
DIRECT=
Y
N
Y
Enables DAR and DDAR.
N
Disables DAR and DDAR.
FILES=
<file-matchingpattern>
None
EXCLUDE=
<file-matchingpattern>
None
RESTORE_HARDLINK
_BY_TABLE=
Y
N
CHECKPOINT_INTERVAL
_IN_BYTES=
262
<size>
5 GB
NDMP backup
Environment variable
Valid values
Default
Description
process failed. The <size>
parameter is the space between
each checkpoint.
Note that this variable can only be
set from the DMA. For example, if
you specify 2 GB, your DMA would
create a checkpoint each time 2
GB of data were backed up.
Restartable backups are
supported only for EMC NetWorker
8.1 and later.
263
NDMP backup
264
CHAPTER 14
File retention with SmartLock
265
SmartLock overview
You can prevent users from modifying and deleting files on an EMC Isilon cluster with the
SmartLock software module. You must activate a SmartLock license on a cluster to
protect data with SmartLock.
With the SmartLock software module, you can create SmartLock directories and commit
files within those directories to a write once read many (WORM) state. You cannot erase
or re-write a file committed to a WORM state. After a file is removed from a WORM state,
you can delete the file. However, you can never modify a file that has been committed to
a WORM state, even after it is removed from a WORM state.
Compliance mode
SmartLock compliance mode enables you to protect your data in compliance with the
regulations defined by U.S. Securities and Exchange Commission rule 17a-4. You can
upgrade a cluster to compliance mode during the initial cluster configuration process,
before you activate the SmartLock license. To upgrade a cluster to SmartLock compliance
mode after the initial cluster configuration process, contact Isilon Technical Support.
If you upgrade a cluster to compliance mode, you will not be able to log in to that cluster
through the root user account. Instead, you can log in to the cluster through the
compliance administrator account that is configured either during initial cluster
configuration or when the cluster is upgraded to compliance mode. If you are logged in
through the compliance administrator account, you can perform administrative tasks
through the sudo command.
SmartLock directories
In a SmartLock directory, you can commit a file to a WORM state manually or you can
configure SmartLock to automatically commit the file. You can create two types of
SmartLock directories: enterprise and compliance. However, you can create compliance
directories only if the cluster has been upgraded to SmartLock compliance mode. Before
you can create SmartLock directories, you must activate a SmartLock license on the
cluster.
If you commit a file to a WORM state in an enterprise directory, the file can never be
modified and cannot be deleted until the retention period passes. However, if you are
logged in through the root user account, you can delete the file before the retention
period passes through the privileged delete feature. The privileged delete feature is not
available for compliance directories. Enterprise directories reference the system clock to
facilitate time-dependent operations, including file retention.
Compliance directories enable you to protect your data in compliance with the
regulations defined by U.S. Securities and Exchange Commission rule 17a-4. If you
commit a file to a WORM state in a compliance directory, the file cannot be modified or
deleted before the specified retention period has expired. You cannot delete committed
files, even if you are logged in to the compliance administrator account. Compliance
directories reference the compliance clock to facilitate time-dependent operations,
including file retention.
You must set the compliance clock before you can create compliance directories. You can
set the compliance clock only once. After you set the compliance clock, you cannot
modify the compliance clock time. The compliance clock is controlled by the compliance
clock daemon. Because a user can disable the compliance clock daemon, it is possible
266
for a user to increase the retention period of WORM committed files in compliance mode.
However, it is not possible to decrease the retention period of a WORM committed file.
Do not configure SmartLock settings for a target SmartLock directory unless you are no
longer replicating data to the directory. Configuring an autocommit time period for a
target SmartLock directory can cause replication jobs to fail. If the target SmartLock
directory commits a file to a WORM state, and the file is modified on the source cluster,
the next replication job will fail because it cannot update the file.
Allowed
Yes
Non-SmartLock
SmartLock enterprise
Yes
Non-SmartLock
SmartLock compliance
No
SmartLock enterprise
Non-SmartLock
SmartLock enterprise
SmartLock enterprise
Yes
SmartLock enterprise
SmartLock compliance
No
SmartLock compliance
Non-SmartLock
No
SmartLock compliance
SmartLock enterprise
No
SmartLock compliance
SmartLock compliance
Yes
If you replicate SmartLock directories to another cluster with SyncIQ, the WORM state of
files is replicated. However, SmartLock directory configuration settings are not transferred
to the target directory.
For example, if you replicate a directory that contains a committed file that is set to expire
on March 4th, the file is still set to expire on March 4th on the target cluster. However, if
267
the directory on the source cluster is set to prevent files from being committed for more
than a year, the target directory is not automatically set to the same restriction.
If you back up data to an NDMP device, all SmartLock metadata relating to the retention
date and commit status is transferred to the NDMP device. If you restore data to a
SmartLock directory on the cluster, the metadata persists on the cluster. However, if the
directory that you restore to is not a SmartLock directory, the metadata is lost. You can
restore to a SmartLock directory only if the directory is empty.
SmartLock considerations
u
It is recommended that you create files outside of SmartLock directories and then
transfer them into a SmartLock directory after you are finished working with the files.
If you are uploading files to a cluster, it is recommended that you upload the files to a
non-SmartLock directory, and then later transfer the files to a SmartLock directory. If a
file is committed to a WORM state while the file is being uploaded, the file will
become trapped in an inconsistent state.
Files can be committed to a WORM state while they are still open. If you specify an
autocommit time period for a directory, the autocommit time period is calculated
according to the length of time since the file was last modified, not when the file was
closed. If you delay writing to an open file for more than the autocommit time period,
the file will be committed to a WORM state the next time you attempt to write to it.
In a Microsoft Windows environment, if you commit a file to a WORM state, you can
no longer modify the hidden or archive attributes of the file. Any attempt to modify
the hidden or archive attributes of a WORM committed file will generate an error. This
can prevent third-party applications from modifying the hidden or archive attributes.
Retention periods
A retention period is the length of time that a file remains in a WORM state before being
released from a WORM state. You can configure SmartLock directory settings that enforce
default, maximum, and minimum retention periods for the directory.
If you manually commit a file, you can optionally specify the date that the file is released
from a WORM state. You can configure a minimum and a maximum retention period for a
SmartLock directory to prevent files from being retained for too long or too short a time
period. It is recommended that you specify a minimum retention period for all SmartLock
directories.
For example, assume that you have a SmartLock directory with a minimum retention
period of two days. At 1:00 PM on Monday, you commit a file to a WORM state, and
specify the file to be released from a WORM state on Tuesday at 3:00 PM. The file will be
released from a WORM state two days later on Wednesday at 1:00 PM, because releasing
the file earlier would violate the minimum retention period.
You can also configure a default retention period that is assigned when you commit a file
without specifying a date to release the file from a WORM state.
269
period, the file is automatically committed to a WORM state the next time that file is
accessed by a user.
After the autocommit time period for a file passes, the file continues to reference the
current autocommit time period until the file is accessed by a user. Therefore, increasing
the autocommit time period of a directory might cause files to be committed to a WORM
state later than expected. For example, assume that you have a SmartLock directory with
an autocommit time period of one day, and an expiration period of one day. You then
copy a file into the SmartLock directory on Monday, at 3:00 PM. At 5:00 PM on Tuesday,
you increase the autocommit time period to two days. If the file was not accessed, users
can modify or delete the file until 3:00 PM on Wednesday.
Decreasing the autocommit time period of a directory can cause a file to be released from
a WORM state earlier than expected. For example, assume that you have a SmartLock
directory with an autocommit time period of one day, and a default expiration period of
one day. You then copy a file into the SmartLock directory on Monday, at 3:00 PM. If, at
4:00 PM on Tuesday, the file was not accessed by a user, and you decrease the
autocommit time period to two hours, the file is set to be removed from a WORM state at
5:00 PM on Tuesday, instead of 3:00 PM on Wednesday.
Modifying the minimum, maximum, or default retention period of a SmartLock directory
can modify the retention period of files, even after the autocommit time period of a file
expires. For example, assume that you have a SmartLock directory with an autocommit
time period of two days, and a default expiration period of one day. You then copy a file
into the SmartLock directory on Monday, at 3:00 PM. If, by 4:00 PM on Wednesday, the
file was not accessed by a user, and you decrease the default retention period to two
hours, the file is removed from a WORM state at 5:00 PM on Wednesday, instead of 3:00
PM on Thursday.
If you specify an autocommit time period along with a minimum, maximum, or default
retention period, the retention period is calculated according to the time that the
autocommit period expires. For example, assume that you have a SmartLock directory
with a minimum retention period of two days and an autocommit time period of one day.
At 1:00 PM on Monday, you modify a file; then, at 5:00 PM on Tuesday, you access the
file, and the file is committed to a WORM state. The retention period expires on Thursday
at 1:00 PM, two days after the autocommit time period for the file expired.
If the autocommit time period expires for a file, and the file is accessed by a user, the file
is committed to a WORM state. However, the read-write permissions of the file are not
modified. The file is still committed to a WORM state; the file can never be modified, and
it cannot be deleted until the specified retention period expires. However, the WORM
state is not indicated by the read-write permissions.
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
270
2. Run the isi worm domains create command to create a SmartLock directory.
The path specified in the isi worm domains create command cannot be the
path of an existing directory.
The following command creates a compliance directory with a default retention period
of four years, a minimum retention period of three years, and an maximum retention
period of five years:
sudo isi worm mkdir --path /ifs/data/dir --compliance \
--default 4y --min 3y --max 5y
It is recommended that you set SmartLock configuration settings only once and do not
modify the settings after files are added to the SmartLock directory.
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
2. Modify SmartLock configuration settings by running the isi worm modify
command.
For example, the following command sets the default retention period to one year:
isi worm modify --path /ifs/data/protected_directory \
--default 1y
271
3. Optional: To view detailed information about a specific SmartLock directory, run the
isi worm domains view command.
The following command displays detailed information about /ifs/data/
SmartLock/directory2:
isi worm domains view /ifs/data/SmartLock/directory2
65537
/ifs/data/SmartLock/directory2
enterprise
4295426060
30m
off
1Y
3M
3/32 Max
272
Specifies years
Specifies months
Specifies weeks
Specifies days
Value Description
y
Specifies years
Specifies months
Specifies weeks
Specifies days
Value Description
y
Specifies years
Specifies months
Specifies weeks
Specifies days
Autocommit period
The autocommit time period for the directory. After a file exists in this SmartLock
directory without being modified for the specified time period, the file is
automatically committed the next time the file is accessed by a user.
Times are expressed in the format "<integer> <time>", where <time> is one of the
following values:
-
Value Description
y
Specifies years
Specifies months
Specifies weeks
Specifies days
Specifies hours
Specifies minutes
273
Privileged delete
Indicates whether files in the directory can be deleted through the privileged delete
functionality.
On
A root user can delete files committed to a WORM state by running the isi
worm filedelete command.
Off
WORM committed files cannot be deleted, even through the isi worm
filedelete command.
Disabled (Permanently)
WORM committed files cannot be deleted, even through the isi worm
filedelete command. After this setting is set, the setting cannot be
modified.
274
3. Specify the name of the file you want to set a retention period for by creating an
object.
The file must exist in a SmartLock directory.
The following command creates an object for /smartlock/file.txt:
$file = Get-Item "\\cluster.ip.address.com\ifs\smartlock\file.txt"
4. Specify the retention period by setting the last access time for the file.
The following command sets an expiration date of July 1, 2015 at 1:00 PM:
$file.LastAccessTime = Get-Date "2015/7/1 1:00 pm"
275
3. Delete the WORM committed file by running the isi worm filedelete command.
For example, the following command deletes /ifs/worm/enterprise/file:
isi worm filedelete /ifs/worm/enterprise/file
276
277
CHAPTER 15
Protection domains
Protection domains
279
Protection domains
280
Copying a large number of files into a protection domain might take a very long time
because each file must be marked individually as belonging to the protection
domain.
You cannot move directories in or out of protection domains. However, you can move
a directory contained in a protection domain to another location within the same
protection domain.
Creating a protection domain for a directory that contains a large number of files will
take more time than creating a protection domain for a directory with fewer files.
Because of this, it is recommended that you create protection domains for directories
while the directories are empty, and then add files to the directory.
Protection domains
281
Protection domains
SmartLock
SmartLock domain of an enterprise directory.
Compliance
SmartLock domain of a compliance directory.
SyncIQ
SyncIQ domain that prevents users from modifying files and directories.
SyncIQ, Writable
SyncIQ domain that allows users to modify files and directories.
SnapRevert
SnapRevert domain that prevents users from modifying files and directories while a
snapshot is being reverted.
Writable, SnapRevert
SnapRevert domain that allows users to modify files and directories.
If Incomplete is appended to a domain type, OneFS is in the process of creating the
domain. An incomplete domain does not prevent files from being modified or deleted.
282
CHAPTER 16
Data-at-rest-encryption
Data-at-rest-encryption
283
Data-at-rest-encryption
Self-encrypting drives
Self-encrypting drives store data on a EMC Isilon cluster that is specially designed for
data-at-rest encryption.
Data-at-rest- encryption on self-encrypted drives occurs when data that is stored on a
device is encrypted to prevent unauthorized data access. All data written to the storage
device is encrypted when it is stored, and all data read from the storage device is
decrypted when it is read. The stored data is encrypted with a 256-bit data AES
encryption key and decrypted in the same manner. OneFS controls data access by
combining the drive authentication key with on-disk data-encryption keys.
Note
All nodes in a cluster must be of the self-encrypting drive type. Mixed nodes are not
supported.
When a drive is smartfailed and removed from a node, the encryption key on the drive
is removed. Because the encryption key for reading data from the drive must be the
same key that was used when the data was written, it is impossible to decrypt data
that was previously written to the drive. When you smartfail and then remove a drive,
it is cryptographically erased.
Note
284
When a self-encrypting drive loses power, the drive locks to prevent unauthorized
access. When power is restored, data is again accessible when the appropriate drive
authentication key is provided.
Data-at-rest-encryption
Before you begin the data-migration process, both clusters must be upgraded to the
same OneFS version.
During data migration, an error is generated that indicates you are running in mixed
mode, which is not supported and is not secure. The data migrated to the self-encrypted
drives is not secure until the smartfail process is completed for the non-encrypted drives.
CAUTION
HEALTHY
Description
-
Interface
CLI, web
administration
interface
SMARTFAIL or
Smartfail or
restripe in
progress
NOT AVAILABLE
CLI, web
administration
interface
Error
state
Note
285
Data-at-rest-encryption
State
-
SUSPENDED
Description
-
Interface
-
CLI, web
administration
interface
NOT IN USE
REPLACE
CLI only
STALLED
CLI only
NEW
USED
CLI only
PREPARING
CLI only
EMPTY
CLI only
WRONG_TYPE
CLI only
BOOT_DRIVE
CLI only
SED_ERROR
CLI , web
administration
interface
Note
ERASE
286
CLI only
Error
state
Data-at-rest-encryption
State
-
INSECURE
Description
-
Interface
CLI only
Error
state
X
Note
UNENCRYPTED
SED
Web
administration
interface only
Note
[SMARTFAIL]
SN:Z296M8HK
[HEALTHY]
SN:Z296M8N5
[HEALTHY]
SN:Z296LBP4
[HEALTHY]
SN:Z296LCJW
[HEALTHY]
SN:Z296M8XB
[HEALTHY]
SN:Z295LXT7
[HEALTHY]
SN:Z296M8ZF
[HEALTHY]
SN:Z296M8SD
[HEALTHY]
SN:Z296M8QA
[HEALTHY]
SN:Z296M8Q7
[HEALTHY]
SN:Z296M8SP
[HEALTHY]
SN:Z296M8QZ
[REPLACE]
SN:Z296M8HK
287
Data-at-rest-encryption
Bay 2
00009330EYE03
Bay 3
00009330EYE03
Bay 4
00009327BYE03
Bay 5
00009330KYE03
Bay 6
000093172YE03
Bay 7
00009330KYE03
Bay 8
00009330EYE03
Bay 9
00009330EYE03
Bay 10
00009330EYE03
Bay 11
00009330EYE04
Bay 12
00009330JYE03
Lnum 10
/dev/da2
Lnum 9
/dev/da3
Lnum 8
/dev/da4
Lnum 7
/dev/da5
Lnum 6
/dev/da6
Lnum 5
/dev/da7
Lnum 4
/dev/da8
Lnum 3
/dev/da9
Lnum 2
/dev/da10
Lnum 1
/dev/da11
Lnum 0
/dev/da12
[HEALTHY]
SN:Z296M8N5
[HEALTHY]
SN:Z296LBP4
[HEALTHY]
SN:Z296LCJW
[HEALTHY]
SN:Z296M8XB
[HEALTHY]
SN:Z295LXT7
[HEALTHY]
SN:Z296M8ZF
[HEALTHY]
SN:Z296M8SD
[HEALTHY]
SN:Z296M8QA
[HEALTHY]
SN:Z296M8Q7
[HEALTHY]
SN:Z296M8SP
[HEALTHY]
SN:Z296M8QZ
[REPLACE]
SN:Z296M8HK
[HEALTHY]
SN:Z296M8N5
[SMARTFAIL]
SN:Z296LBP4
[HEALTHY]
SN:Z296LCJW
[HEALTHY]
SN:Z296M8XB
[HEALTHY]
SN:Z295LXT7
[HEALTHY]
SN:Z296M8ZF
[HEALTHY]
SN:Z296M8SD
[HEALTHY]
SN:Z296M8QA
[HEALTHY]
SN:Z296M8Q7
[HEALTHY]
SN:Z296M8SP
[HEALTHY]
SN:Z296M8QZ
Smartfail is unsuccessful for bay 3, changing the drive state to ERASE. Because the drive
could not be crypto-erased, OneFS attempts to delete the drive password:
my-cluster# isi dev
Node 1, [ATTN]
Bay 1
Lnum 11
000093172YE04 /dev/da1
Bay 2
Lnum 10
00009330EYE03 /dev/da2
Bay 3
Lnum 9
00009330EYE03 /dev/da3
Bay 4
Lnum 8
00009327BYE03 /dev/da4
Bay 5
Lnum 7
00009330KYE03 /dev/da5
Bay 6
Lnum 6
000093172YE03 /dev/da6
Bay 7
Lnum 5
288
[REPLACE]
SN:Z296M8HK
[HEALTHY]
SN:Z296M8N5
[ERASE]
SN:Z296LBP4
[HEALTHY]
SN:Z296LCJW
[HEALTHY]
SN:Z296M8XB
[HEALTHY]
SN:Z295LXT7
[HEALTHY]
SN:Z296M8ZF
Data-at-rest-encryption
00009330KYE03
Bay 8
00009330EYE03
Bay 9
00009330EYE03
Bay 10
00009330EYE03
Bay 11
00009330EYE04
Bay 12
00009330JYE03
/dev/da7
Lnum 4
/dev/da8
Lnum 3
/dev/da9
Lnum 2
/dev/da10
Lnum 1
/dev/da11
Lnum 0
/dev/da12
[HEALTHY]
SN:Z296M8SD
[HEALTHY]
SN:Z296M8QA
[HEALTHY]
SN:Z296M8Q7
[HEALTHY]
SN:Z296M8SP
[HEALTHY]
SN:Z296M8QZ
289
CHAPTER 17
SmartQuotas
SmartQuotas
291
SmartQuotas
SmartQuotas overview
The SmartQuotas module is an optional quota-management tool that monitors and
enforces administrator-defined storage limits. Using accounting and enforcement quota
limits, reporting capabilities, and automated notifications, SmartQuotas manages
storage use, monitors disk storage, and issues alerts when disk-storage limits are
exceeded.
Quotas help you manage storage usage according to criteria that you define. Quotas are
used as a method of trackingand sometimes limitingthe amount of storage that a
user, group, or project consumes. Quotas are a useful way of ensuring that a user or
department does not infringe on the storage that is allocated to other users or
departments. In some quota implementations, writes beyond the defined space are
denied, and in other cases, a simple notification is sent.
The SmartQuotas module requires a separate license. For additional information about
the SmartQuotas module or to activate the module, contact your EMC Isilon sales
representative.
Quota types
OneFS uses the concept of quota types as the fundamental organizational unit of storage
quotas. Storage quotas comprise a set of resources and an accounting of each resource
type for that set. Storage quotas are also called storage domains.
Storage quotas creation requires three identifiers:
u
292
SmartQuotas
Note
You should not create quotas of any type on the OneFS root (/ifs). A root-level quota
may significantly degrade performance.
In this example, the default-user type created a new specific-user type automatically
(user:admin) and added the new usage to it. Default-user does not have any usage
because it is used only to generate new quotas automatically. Default-user enforcement
is copied to a specific-user (user:admin), and the inherited quota is called a linked quota.
In this way, each user account gets its own usage accounting.
Defaults can overlap. For example, default-user@/ifs/dir-1 and default-user@/ifs/cs
both may be defined. If the default enforcement changes, OneFS storage quotas
propagate the changes to the linked quotas asynchronously. Because the update is
asynchronous, there is some delay before updates are in effect. If a default type, such as
every user or every group, is deleted, OneFS deletes all children that are marked as
inherited. As an option, you can delete the default without deleting the children, but it is
important to note that this action breaks inheritance on all inherited children.
Continuing with the example, add another file that is owned by the root user. Because the
root type exists, the new usage is added to it.
my-OneFS-1# touch /ifs/dir-1/anotherfile
my-OneFS-1# isi quota ls -v --path=/ifs/dir-1 --format=list
Type: default-user
AppliesTo: DEFAULT
Path: /ifs/dir-1
Snap: No
Thresholds
Hard : Default quota type
293
SmartQuotas
Configuration changes for linked quotas must be made on the parent quota that the
linked quota is inheriting from. Changes to the parent quota are propagated to all
children. To override configuration from the parent quota, you must unlink the quota first.
SmartQuotas
Accounting
The accounting option tracks but does not limit disk-storage use. Using the
accounting option for a quota, you can monitor inode count and physical and logical
space resources. Physical space refers to all of the space used to store files and
directories, including data and metadata in the domain. Logical space refers to the
sum of all files sizes, excluding file metadata and sparse regions. User data storage
is tracked using logical-space calculations, which do not include protection
overhead. As an example, by using the accounting option, you can do the following:
u
Track the amount of disk space used by various users or groups to bill each user,
group, or directory for only the disk space used.
Review and analyze reports that help you identify storage usage patterns and
define storage policies.
Plan for capacity and other storage needs.
Enforcement limits
Enforcement limits include all of the functionality of the accounting option, plus the
ability to limit disk storage and send notifications. Using enforcement limits, you can
logically partition a cluster to control or restrict how much storage that a user, group,
or directory can use. For example, you can set hard- or soft-capacity limits to ensure
that adequate space is always available for key projects and critical applications and
to ensure that users of the cluster do not exceed their allotted storage capacity.
Optionally, you can deliver real-time email quota notifications to users, group
managers, or administrators when they are approaching or have exceeded a quota
limit.
Note
If a quota type uses the accounting-only option, enforcement limits cannot be used for
that quota.
The actions of an administrator logged in as root may push a domain over a quota
threshold. For example, changing the protection level or taking a snapshot has the
potential to exceed quota parameters. System actions such as repairs also may push a
quota domain over the limit.
The system provides three types of administrator-defined enforcement thresholds.
Threshold
type
-
Hard
Description
-
Limits disk usage to a size that cannot be exceeded. If an operation, such as a file
write, causes a quota target to exceed a hard quota, the following events occur:
l
Allows a limit with a grace period that can be exceeded until the grace period
expires. When a soft quota is exceeded, an alert is logged to the cluster and a
notification is issued to specified recipients; however, data writes are permitted
during the grace period.
295
SmartQuotas
Threshold
type
-
Description
-
If the soft threshold is still exceeded when the grace period expires, data writes
fail, and a hard-limit notification is issued to the recipients you have specified.
Writes resume when the usage falls below the threshold.
Advisory
Disk-usage calculations
For each quota that you configure, you can specify whether data-protection overhead is
included in future disk-usage calculations.
Most quota configurations do not need to include overhead calculations. If you do not
include data-protection overhead in usage calculations for a quota, future disk-usage
calculations for the quota include only the space that is required to store files and
directories. Space that is required for the data-protection setting of the cluster is not
included.
Consider the same example user, who is now restricted by a 40 GB quota that does not
include data-protection overhead in its disk-usage calculations. If your cluster is
configured with a 2x data-protection level and the user writes a 10 GB file to the cluster,
that file consumes 20 GB of space but the 10GB for the data-protection overhead is not
counted in the quota calculation. In this example, the user has reached 25 percent of the
40 GB quota by writing a 10 GB file to the cluster. This method of disk-usage calculation
is recommended for most quota configurations.
If you include data-protection overhead in usage calculations for a quota, future diskusage calculations for the quota include the total amount of space that is required to
store files and directories, in addition to any space that is required to accommodate your
data-protection settings, such as parity or mirroring. For example, consider a user who is
restricted by a 40 GB quota that includes data-protection overhead in its disk-usage
calculations. If your cluster is configured with a 2x data-protection level (mirrored) and
the user writes a 10 GB file to the cluster, that file actually consumes 20 GB of space: 10
GB for the file and 10 GB for the data-protection overhead. In this example, the user has
reached 50 percent of the 40 GB quota by writing a 10 GB file to the cluster.
Note
Cloned and deduplicated files are treated as ordinary files by quotas. If the quota
includes data protection overhead, the data protection overhead for shared data is not
included in the usage calculation.
You can configure quotas to include the space that is consumed by snapshots. A single
path can have two quotas applied to it: one without snapshot usage, which is the default,
and one with snapshot usage. If you include snapshots in the quota, more files are
included in the calculation than are in the current directory. The actual disk usage is the
sum of the current directory and any snapshots of that directory. You can see which
snapshots are included in the calculation by examining the .snapshot directory for the
quota path.
296
SmartQuotas
Note
Only snapshots created after the QuotaScan job finishes are included in the calculation.
Quota notifications
Quota notifications are generated for enforcement quotas, providing users with
information when a quota violation occurs. Reminders are sent periodically while the
condition persists.
Each notification rule defines the condition that is to be enforced and the action that is to
be executed when the condition is true. An enforcement quota can define multiple
notification rules. When thresholds are exceeded, automatic email notifications can be
sent to specified users, or you can monitor notifications as system alerts or receive
emails for these events.
Notifications can be configured globally, to apply to all quota domains, or be configured
for specific quota domains.
Enforcement quotas support the following notification settings. A given quota can use
only one of these settings.
Limit notification settings
-
Description
-
297
SmartQuotas
Instant notifications
Includes the write-denied notification, triggered when a hard threshold denies a
write, and the threshold-exceeded notification, triggered at the moment a hard, soft,
or advisory threshold is exceeded. These are one-time notifications because they
represent a discrete event in time.
Ongoing notifications
Generated on a scheduled basis to indicate a persisting condition, such as a hard,
soft, or advisory threshold being over a limit or a soft threshold's grace period being
expired for a prolonged period.
Quota reports
The OneFS SmartQuotas module provides reporting options that enable administrators to
manage cluster resources and analyze usage statistics.
Storage quota reports provide a summarized view of the past or present state of the
quota domains. After raw reporting data is collected by OneFS, you can produce data
summaries by using a set of filtering parameters and sort types. Storage-quota reports
include information about violators, grouped by threshold types. You can generate
reports from a historical data sample or from current data. In either case, the reports are
views of usage data at a given time. OneFS does not provide reports on data aggregated
over time, such as trending reports, but you can use raw data to analyze trends. There is
no configuration limit on the number of reports other than the space needed to store
them.
OneFS provides three methods of data collection and reporting:
u
Ad hoc reports are generated and saved at the request of the user.
Creating quotas
You can create two types of storage quotas to monitor data: accounting quotas and
enforcement quotas. Storage quota limits and restrictions can apply to specific users,
groups, or directories.
The type of quota that you create depends on your goal.
u
Enforcement quotas monitor and limit disk usage. You can create enforcement
quotas that use any combination of hard limits, soft limits, and advisory limits.
Note
298
SmartQuotas
Note
After you create a new quota, it begins to report data almost immediately, but the data is
not valid until the QuotaScan job completes. Before using quota data for analysis or other
purposes, verify that the QuotaScan job has finished.
To include snapshot data in the accounting quota, select the Include Snapshot
Data check box.
To include snapshot data in the accounting quota, select the Include Snapshot
Data check box.
299
SmartQuotas
5. In the Directory path field, type the path and directory for the quota, or click Browse,
and then select a directory.
6. Optional: In the Usage Accounting area, click the Include Snapshot Data check box,
the Include Data-Protection Overhead check box, or both to include them in the
quota.
7. In the Usage Limits area, click Specify Usage Limits.
8. Click the check box next to the option for each type of limit that you want to enforce.
9. Type numerals in the fields and select from the lists the values that you want to use
for the quota.
10.In the Limit Notations area, click the notification option that you want to apply to the
quota.
11.To generate an event notification, select the Create cluster event check box.
12.Optional: If you selected the option to use custom notification rules, click the link to
expand the custom notification type that applies to the usage-limit selections.
13.Click Create Quota.
After you finish
After you create a quota, it begins to report data almost immediately but the data is not
valid until the QuotaScan job completes. Before using quota data for analysis or other
purposes, verify that the QuotaScan job has finished.
Managing quotas
You can modify the configured values of a storage quota, and you can enable or disable a
quota. You can also create quota limits and restrictions that apply to specific users,
groups, or directories.
Quota management in OneFS is simplified by the quota search feature, which helps you
to locate a quota or quotas by using filters. You can unlink quotas that are associated
with a parent quota, and configure custom notifications for quotas. You can also disable
a quota temporarily and then enable it when needed.
Note
SmartQuotas
To search for only default users, select the Only show default users checkbox.
To search for only default groups, select the Only show default groups check box.
To search for only quotas that are in violations, select the Only show quotas for
which usage limits are currently in violation check box.
To clear the result set and display all storage quotas, in the Quotas & Usage area, select
Show all quotas and usage for this report for Report Filters, and then click Update
Display.
Manage quotas
Quotas help you monitor and analyze the current or historical use of disk storage. You
can search for quotas, and you can view, modify, delete, and unlink a quota.
An initial QuotaScan job must run for the default or scheduled quotas, or the data
displayed may be incomplete.
Before you modify a quota, consider how the changes will affect the file system and end
users.
Note
u
The options to edit or delete a quota appear only when the quota is not linked to a
default quota.
The option to unlink a quota is available only when the quota is linked to a default
quota.
Procedure
1. Click File System Management > SmartQuotas > Quotas & Usage.
2. From the Quota Report options, select the type of quota report that you want to view
or manage.
l
To monitor and analyze current disk storage use, click Show current quotas and
usage (Live Report).
To monitor and analyze historical disk storage use, click Show archived quota
report to select from the list of archived scheduled and manually generated quota
reports.
3. For Report Filters, select the filters to be used for this quota report.
l
To view all information in the quota report, click Show all quotas and usage for
this report.
Manage quotas
301
SmartQuotas
To filter the quota report, click Search for specific quotas within this report, and
then select the filters that you want to apply.
Configuration changes for linked quotas must be made on the parent (default)
quota that the linked quota is inheriting from. Changes to the parent quota are
propagated to all children. If you want to override configuration from the parent
quota, you must first unlink the quota.
302
SmartQuotas
Threshold exceeded
Over-quota reminder
To configure the number of live reports that you want to archive, type the number
of reports in the Limit archive size field.
To specify an archive directory that is different from the default, in the Archive
Directory field, type the path or click Browse to select the path.
6. In the Manual Report Archiving area, you can configure the following size and
directory options:
l
To configure the number of live reports that you want to archive, type the number
of reports in the Limit archive size field.
To specify an archive directory that is different from the default, in the Archive
Directory field, type the path or click Browse to select the path.
7. In the Email Mapping Rules area, choose each mapping rule that you want to use by
selecting the check box in the Provider Type column.
8. In the Notification Rules area, define default notification rules for each rule type.
l
To expand the list of limit notifications rules types, click Default Notifications
Settings.
Managing quota notifications
303
SmartQuotas
To set the advisory-limit options that you want, click Event: Advisory Limit Value
Exceeded and Event: While Advisory Limit Remains Exceeded.
To display default settings for soft-limit notifications, click Soft Limit Notification
Rules.
To set the soft-limit options that you want, click Event: Soft Limit Value Exceeded,
Event: While Soft Limit Remains Exceeded, Event: Soft Limit Grace Period Expired,
and Event: Soft Limit Write Access Denied.
To display the options for a hard-limit notification rule, click Hard Limit
Notification Rules .
To set the hard-limit options that you want, click Event: Hard Limit Write Access
Denied and Event: While Hard Limit Remains Exceeded.
9. Click Save.
After you finish
After you create a new quota, it begins to report data almost immediately, but the data is
not valid until the QuotaScan job completes. Before using quota data for analysis or other
purposes, verify that the QuotaScan job has finished.
304
SmartQuotas
You must be logged in to the web administration interface to perform this task.
Procedure
1. Click File System Management > SmartQuotas > Settings.
2. Optional: In the Email Mapping area, click Create an email mapping rule.
3. From the Provider Type list, select the provider type for this notification rule.
4. From the Current Domain list, select the domain that you want to use for the mapping
rule.
5. In the Map-to-Domain field, type the name of the domain that you want to map email
notifications to.
Repeat this step if you want to map more than one domain.
6. Click Save Rule.
The following example illustrates a custom email template to notify recipients about an
exceeded quota.
Text-file contents with variables
The disk quota on directory <ISI_QUOTA_PATH> owned by
<ISI_QUOTA_OWNER> was exceeded.
The <ISI_QUOTA_TYPE> quota limit is <ISI_QUOTA_THRESHOLD>, and
<ISI_QUOTA_USAGE> is in use. Please free some disk space
Map an email notification rule for a quota
305
SmartQuotas
306
SmartQuotas
Description
ls -a *.xml
ls <filename>.xml
Directory Path
Description
-
307
SmartQuotas
Option
-
User Quota
Description
-
Group Quota
No Usage Limit
Description
Send email
Exceeded Remains
exceeded
Yes
Yes
Notify owner
Yes
Yes
Notify another
Yes
Yes
Message template
Select from the following template types for use in formatting email
notifications:
Yes
Yes
Create cluster event Select to generate an event notification for the quota when exceeded.
Yes
Yes
Delay
Specify the length of time (hours, days, weeks) to delay before generating a
notification.
Yes
No
Frequency
Specify the notification and alert frequency: daily, weekly, monthly, yearly;
depending on selection, specify intervals, day to send, time of day, multiple
emails per rule.
No
Yes
308
Custom
SmartQuotas
Send email
Description
Exceeded Remains
exceeded
Yes
Yes
Grace
period
expired
-
Yes
Write
access
denied
-
Yes
Notify owner
Yes
Yes
Yes
Yes
Notify another
Yes
Yes
Yes
Yes
Message
template
Yes
Yes
Yes
Yes
Custom
Create cluster
event
Yes
Yes
Yes
Yes
Delay
Yes
No
No
Yes
Frequency
No
Yes
Yes
No
Send email
Description
-
Write access
denied
-
Yes
Exceeded
-
Yes
Notify owner
Yes
Yes
Notify another
Yes
Yes
Message template
Select from the following template types for use in formatting email
notifications:
Yes
Yes
309
SmartQuotas
Option
Description
Write access
denied
Exceeded
-
Custom
Create cluster event Select to generate an event notification for the quota when exceeded.
Yes
Yes
Delay
Specify the length of time (hours, days, weeks) to delay before generating a
notification.
Yes
No
Frequency
Specify the notification and alert frequency: daily, weekly, monthly, yearly;
depending on selection, specify intervals, day to send, time of day, multiple
emails per rule.
No
Yes
Notification setting
Use Default Notification Rules
Description
Uses the default notification rules that you configured for
the specified threshold type.
Turn Off Notifications for this Quota Disables all notifications for the quota.
Use Custom Notification Rules
Description
Scheduled
reporting
Report
frequency
On. Reports run automatically according to the schedule that you specify.
Specifies the interval for this report to run: daily, weekly, monthly, or yearly. You can use the following options to
further refine the report schedule.
Generate report every. Specify the numeric value for the selected report frequency; for example, every 2
months.
Generate reports on. Select the day or multiple days to generate reports.
Select report day by. Specify date or day of the week to generate the report.
310
SmartQuotas
Setting
-
Description
-
Generate one report per specified by. Set the time of day to generate this report.
Generate multiple reports per specified day. Set the intervals and times of day to generate the report for
that day.
Scheduled
Determines the maximum number of scheduled reports that are available for viewing on the SmartQuotas
report archiving Reports page.
Limit archive size for scheduled reports to a specified number of reports. Type the integer to specify the
maximum number of reports to keep.
Archive Directory. Browse to the directory where you want to store quota reports for archiving.
Manual report
archiving
Determines the maximum number of manually generated (on-demand) reports that are available for viewing on
the SmartQuotas Reports page.
Limit archive size for live reports to a specified number of reports. Type the integer to specify the maximum
number of reports to keep.
Archive Directory. Browse to the directory where you want to store quota reports for archiving.
ISI_QUOTA_PATH
Description
-
Example
-
/ifs/data
20 GB
ISI_QUOTA_USAGE
10.5 GB
ISI_QUOTA_OWNER
jsmith
ISI_QUOTA_TYPE
Threshold type
Advisory
ISI_QUOTA_GRACE
5 days
ISI_QUOTA_NODE
someHost-prod-wf-1
311
CHAPTER 18
Storage Pools
Storage Pools
313
Storage Pools
Feature
Inactive Active
-
Yes
Directed spillover
No
Yes
No
Yes
Yes
Yes
314
Storage Pools
another file pool policy to evaluate the last accessed date, allowing you to target storage
pools best suited for archiving for historical or regulatory purposes.
When you install OneFS, the system creates one file pool for the Isilon cluster. This
default file pool contains all the files in the cluster and is governed by a default file pool
policy. The default file pool policy operations apply to all files not governed by a higherpriority file pool policy. If you activate a SmartPools license, you can create multiple file
pools.
OneFS includes the following basic features.
Default file pool
A single set of files of all types that is governed by the default file pool policy.
Node pools
Groups of equivalence-class nodes that are associated in a single pool of storage.
Tiers
Groups of node pools used to optimize data storage according to your storage
needs.
Activating a SmartPools license adds the following features.
File pools
Logical sets of files that you define by characteristics such as file type, directory, or
age.
File pool policies
Rules-based filtering and operations that you configure to store data in specified
storage pools. By creating file-filtering rules such as file size, type, access time, and
location to configure a file pool policy, you can configure requested protection and
I/O optimization settings and automate data storage according to your needs. If you
activate a SmartPools license, you can create customizable file pool templates that
are optimized for archiving, extra protection, performance, and VMware files.
Storage pool spillover
Automated node-capacity overflow management. Spillover defines how to handle
write operations when a storage pool is not writable. When spillover is enabled, data
is redirected to a specified storage pool. If spillover is disabled, new data writes fail
and an error message appears.
Note
If a SmartPools license has not been activated, files are stored on any available node
pools across the cluster.
Autoprovisioning
Autoprovisioning is the process of automatically assigning storage by node type to
improve the performance and reliability of the file storage system.
When you configure a cluster, OneFS automatically assigns nodes to node pools in your
cluster. This node assignment is called autoprovisioning. Autoprovisioning reduces the
time required for the manual management tasks associated with configuring storage
pools and resource planning.
Autoprovisioning
315
Storage Pools
Note
Nodes are not provisioned, meaning they are not associated with each other and not
writable, until at least three nodes of an equivalence class are added to the cluster. If you
have added only two nodes of an equivalence class to your cluster, no data is stored on
the nodes until you add a third node of the same equivalence-class.
If you remove nodes from a provisioned cluster so that fewer than three equivalenceclass nodes remain, the pool is underprovisioned. In this situation, when two like nodes
remain, they are still writable; if only one node remains, it is not writable but it remains
readable.
The larger number of the two factors (minimum number of virtual drives or percentage
of total disk space), rather than their sum, determines the space allocated for virtual
hot spare.
It is important to understand the following information when configuring VHS settings:
u
If you configure both settings, the enforced minimum value satisfies both
requirements.
u
If you select the option to reduce the amount of available space, free-space
calculations do not include the space reserved for the virtual hot spare. The reserved
virtual hot spare free space is used for write operations unless you select the option
to deny new data writes. If Reduce amount of available space is enabled while Deny
new data writes is disabled, it is possible for the file system to report utilization as
more than 100 percent.
Note
Virtual hot spare reservations affect spillover. If the virtual hot spare option Deny writes
is enabled but Reduce amount of available space is disabled, spillover occurs before the
file system reports 100% utilization.
Spillover
If you activate a SmartPools license, you can designate a storage pool to receive spill
data when the hardware specified by a file pool policy is not writable. If you do not want
316
Storage Pools
data to spill over from a different location because the specified node pool or tier is full or
not writable, you can disable this feature.
Spillover management is available after you activate a SmartPools license. You can direct
write operations to a specified storage pool in the cluster when there is not enough space
to write a file according to the storage pool policy.
Note
Virtual hot spare reservations affect spillover. If the setting Deny writes is enabled but
Deduce amount of available space is disabled, spillover occurs before the file system
reports 100% utilization.
Node pools
A node pool is a grouping of equivalence-class nodes. As you add nodes to the OneFS
cluster, OneFS automatically place them into pools. This is referred to as
autoprovisioning.
Each node in the OneFS cluster is a peer, and any node can handle a data request. You
can apply file pool policies to move files to specific node pools or tiers that have different
performance and capacity characteristics. Each node added to a cluster increases the
cluster's aggregate disk, cache, CPU, and network capacity. OneFS automatically adds
nodes to specific node pools based on matching characteristics such as drive size, RAM,
series, and SSD-node ratio. Nodes with similar characteristics are called equivalenceclass nodes.
Note
Manually managed node pools may not provide the same performance and efficiency as
automatically managed node pools, particularly if your changes result in fewer than 20
nodes in the manually managed node pool. For this reason, it is best to allow OneFS to
autoprovision nodes unless you have an advanced understanding of how the
SmartPools feature works.
Node pools
317
Storage Pools
SSD pools
OneFS clusters can contain both HDDs and SSDs. When OneFS autoprovisions nodes, it
places nodes with SSDs into equivalence-class node pools. The SSD strategy defined in
the default file pool determines how SSD nodes are used within the cluster.
Clusters that include both hard-disk drives (HDDs) and solid-state drives (SSDs) are
optimized by your SSD strategy options to increase performance across a wide range of
workflows. You can configure file pool policies to apply specific SSD strategies as
needed. When you select SSD options during the creation of a file pool policy, you can
identify the directories and files in the OneFS cluster that require faster or slower
performance. When the file pool policy runs, OneFS automatically moves that data to the
appropriate storage pool and drive type.
Global namespace acceleration (GNA) allows data stored on node pools without SSDs to
use SSDs elsewhere in the cluster to store extra metadata mirrors. Extra metadata mirrors
accelerate metadata read operations. You can only enable GNA if 20% or more of the
nodes in the cluster contain at least one SSD and 1.5% or more of the total cluster
storage is SSD-based. For best results, ensure that at least 2.0% of the total cluster
storage is SSD-based before enabling global namespace acceleration.
Note
GNA requires a minimum of 20% of accessible cluster nodes to have SSD drives. If the
ratio of SSD to non-SSD nodes falls below this threshold, GNA is not active even if
enabled. GNA is reactivated when the ratio is corrected. When GNA is inactive, existing
SSD mirrors are readable but newly written metadata does not include the extra SSD
mirror.
The following SSD strategy options are listed in order of slowest to fastest choices:
Avoid SSDs
Writes all associated file data and metadata to HDDs only.
CAUTION
Use this option to free SSD space only after consulting with Isilon Technical Support
personnel. Using this strategy may negatively affect performance.
Metadata read acceleration
Writes both file data and metadata to HDDs. This is the default setting. An extra
mirror of the file metadata is written to SSDs, if available. The SSD mirror is in
addition to the number required to satisfy the requested protection. Enabling GNA
makes read acceleration available to files in node pools that do not contain SSDs.
GNA is only for metadata and extra mirrors.
Metadata read/write acceleration
Writes file data to HDDs and metadata to SSDs, when available. This strategy
accelerates metadata writes in addition to reads but requires about four to five times
more SSD storage than the Metadata read acceleration setting. Enabling GNA does
not affect read/write acceleration.
318
Storage Pools
Data on SSDs
Uses SSD node pools for both data and metadata, regardless of whether global
namespace acceleration is enabled. This SSD strategy does not result in the creation
of additional mirrors beyond the normal requested protection but requires
significantly increased storage requirements compared with the other SSD strategy
options.
Tiers
A tier is a user-defined collection of node pools that you can specify as a storage pool for
files. A node pool can belong to only one tier.
You can create tiers to assign your data to any of the node pools in the tier. For example,
you can assign a collection of node pools to a tier specifically created to store data that
requires high availability and fast access. In a three-tier system, this classification may
be Tier 1. You can classify data that is used less frequently or that is accessed by fewer
users as Tier-2 data. Tier 3 usually comprises data that is seldom used and can be
archived for historical or regulatory purposes.
File pools
File pools are sets of files that you define to apply policy-based control of the storage
characteristics of your data.
The initial installation of OneFS places all files in the cluster into a single file pool, which
is subject to the default file pool policy. SmartPools enables you to define additional file
pools, and create policies that move files in these pools to specific node pools and tiers.
File pool policies match specific file characteristics (such as file size, type, date of last
access or a combination of these and other factors), and define specific storage
operations for files that match them. The following examples demonstrate a few ways you
can configure file pool policies:
u
You can create a file pool policy for a specific file extension that requires high
availability.
u
You can configure a file pool policy to store that type of data in a storage pool that
provides the fastest reads or read/writes.
u
You can create another file pool policy to evaluate last accessed date, allowing you to
store older files in storage pool best suited for archiving for historical or regulatory
purposes.
319
Storage Pools
The default file pool contains all files and is stored in any node pool. Default file pool
operations are defined by settings of the default file pool policy.
You cannot reorder or remove the default file pool policy. The settings in the default file
pool policy apply to all files that are not covered by another file pool policy. For example,
data that is not covered by a file pool policy can be moved to a tier that you identify as a
default for this purpose.
All file pool policy operations are executed when the SmartPools job runs. When new files
are created, OneFS temporarily chooses a storage pool policy, using a mechanism based
on file pool policies used when the last SmartPools job ran. The system may apply new
storage settings and move these files again when the next SmartPools job runs, based on
a matching file pool policy.
To add a node pool that is currently in another tier, expand that tier and drag and
drop the node pool to the target tier name.
To remove a node pool from a tier, drag and drop the node pool to the cluster icon.
3. Continue dragging and dropping node pools until you complete the tier.
Each node that you added to the tier appears under the tier name when it is in an
expanded state.
320
Storage Pools
Managing tiers
You can group node pools into tiers and move node pools among tiers to most efficiently
use resources or for other cluster management purposes.
Create a tier
You can group node pools into a tier that you can specify as a storage pool for files.
Procedure
1. Click File System Management > SmartPools > Summary.
The SmartPools page appears and displays two groupings: the current capacity usage
and a list of tiers and node pools.
2. In the Tiers & Node Pools section, click Create a Tier.
3. In the dialog box that displays, enter a name for the tier, and click Submit.
The tier appears in the list of tiers and node pools.
4. Select and drag a node pool to the tier name to add it to the tier. Continue dragging
and dropping node pools until you complete the tiered group.
Each node pool that you added to the tier appears under the tier name when it is in an
expanded state.
Rename a tier
You can modify the name of a tier.
A tier name can contain alphanumeric characters and underscores but cannot begin with
a number.
Procedure
1. Click File System Management > SmartPools > Summary.
The SmartPools page appears and displays two groupings: the current capacity usage
and a list of tiers and node pools.
2. In the Tiers & Node Pools area, in the row of the tier you want to rename, click Edit.
3. In the dialog box that displays, type a name for this tier and click Submit.
The newly named tier appears in the list of tiers and node pools.
Delete a tier
When you delete a tier, the nodes it contains become top-level storage pools.
Procedure
1. Click File System Management > SmartPools > Summary.
The SmartPools page appears and displays two groupings: current capacity usage and
a list of tiers and node pools.
2. In the Tiers & Node Pools area, in the row of the tier that you want to delete, click
Delete.
3. In the confirmation dialog box that displays, click Yes to confirm the deletion.
Managing tiers
321
Storage Pools
Results
The tier is removed from the list of tiers and node pools.
You can configure each of these settings for a file in four separate file pool policies,
include them all in one file pool policy, or use a combination of the two. When the
SmartPools job runs, it applies file pool policies in the order of the display. When a file
type matches the criteria defined in the filter settings, the operations are applied. After
the list is traversed, if any of the four operations are not applied to a file type, the
operations defined in the default file pool policy then apply to those files. In this way, the
default file pool policy ensures that all four operations apply to each file, even if an
operation is missing from a user-defined file pool policy.
If a file belongs to multiple file pools with operations specifying the same setting, only
the operation of the file pool policy with the higher priority applies. Consider the
following example: File pool policy 1 targets the X storage pool for all JPG files, and file
pool policy 2 targets the NL storage pool for all files created more than month ago.
Sometimes files match the criteria in both file pool policies, so those files are stored in
the X storage pool because that file pool policy is higher in the list.
If a file type matches multiple file pool policies, subsequent file pool policies in the list
are not evaluated. If one file pool policy operation sets the data storage target of JPG files
to a nearline node pool and a subsequent file pool policy operation sets the data access
pattern to random for all files smaller than 2 MB, then all JPG files smaller than 2 MB are
moved to nearline storage, but their data access pattern is set to random only if the Stop
processing option is not selected in the first file pool policy. If that option is selected,
the data access pattern for JPG files smaller than 2MB is set according to the operation
defined in the default file pool policy.
OneFS provides customizable template policies that archive older files, increase the
requested protection for specified files, send files that are saved to a particular path to a
higher-performance storage pool, and change the data access pattern for VMWare files.
322
Storage Pools
You also can copy any file pool policy except the default file pool policy, and modify the
settings to meet your needs.
Note
You can reorder the file pool policy list at any time, but the default file pool policy is
always last in the list of file pool policies.
If existing file pool policies direct data to a specific storage pool, do not configure other
file pool policies that match this data with anywhere for the Data storage target option.
Because the specified storage pool is included when you use anywhere, target specific
storage pools to avoid unintentional file storage locations.
Procedure
1. Click File System Management > SmartPools > File Pool Policies.
The SmartPools page appears and displays three groupings: a list of file pool policies,
a list of template policies, and latest scan job results.
2. On the SmartPools page, click Add a file pool policy.
3. In the Basic Settings area, type a policy name and a description for this file pool
policy.
4. In the Filter Settings areas, click Add criteria.
5. In the Configure File Matching Criteria window, configure file matching criteria and
click Add.
6. Optional: Select or clear the Stop processing more rules for files matching this filter
check box.
Note
If the stop processing option is selected, only the first matching policy in the list will
be applied if multiple file pool policies are configured for this entity.
7. In the Protections Settings area, choose and configure the target and requested
protection.
8. In the I/O Optimization Settings area, choose the options that you want this file pool
policy to apply.
9. Click Submit.
The Storage Pools page appears and the file pool policy that you created is displayed
in the list.
Results
The file pool policy is not applied until the next scheduled SmartPools job runs. To run
the job immediately, click Start SmartPools Job.
323
Storage Pools
OneFS supports UNIX shell-style (glob) pattern matching for file name attributes and
paths. For information about the characters that can be selected for pattern matching,
see Valid wildcard characters.
The following table lists the file attributes that you can use to define a file pool policy.
File attribute
-
File name
Specifies
-
Location (Path)
File type
File size
Includes or excludes files based on one of the following filesystem object types:
l
Regular file
Directory
Other
324
Storage Pools
File attribute
-
Specifies
-
You can specify a relative date and time, such as "older than 2
weeks," or a specific date and time, such as "before January 1,
2012." Time settings are based on a 24-hour clock.
Includes or excludes files based on when the file was last
accessed based on the following units of time:
You can specify a relative date and time, such as "older than 2
weeks," or a specific date and time, such as "before January 1,
2012." Time settings are based on a 24-hour clock.
Note
Wildcard Description
-
[a-z ]
325
Storage Pools
Data
storage
target
Description
-
Notes
Note
Data on SSDs
Use SSDs for both data and metadata.
Regardless of whether global
namespace acceleration is enabled, any
SSD blocks reside on the storage target
if there is room.
326
Storage Pools
Setting
-
Snapshot
storage
target
Requested
protection
Description
-
Notes
SmartCache
Data access
pattern
Description
-
Notes
Enables or disables
SmartCache.
327
Storage Pools
If existing file pool policies direct data to a specific storage pool, do not add or modify a
file pool policy to the anywhere option for the Data storage target option. Target a
specific file pool instead.
Procedure
1. Click File System Management > SmartPools > Settings.
2. In the SmartPools Settings section, choose the settings that you want apply as the
global default for Data storage target, Snapshot storage target, or Protection level.
3. Click Submit.
The settings that you selected are applied to any entity that is not covered by another
file pool policy.
328
Storage Pools
If existing file pool policies direct data to a specific storage pool, do not configure other
file pool policies with anywhere for the Data storage target option. Because the
specified storage pool is included when you use anywhere, target specific storage
pools to avoid unintentional file storage locations.
Procedure
1. Click File System Management > SmartPools > File Pool Policies.
2. In the File Pool Policies area, in the Actions column of the file pool policy you want to
modify, click Copy.
3. Make your changes in the appropriate areas and click Submit.
Results
Changes to the file pool policy are applied when the next SmartPools job runs. To run the
job immediately, click Start SmartPools Job.
329
Storage Pools
Procedure
1. Click File System Management > SmartPools > File Pool Policies.
The SmartPools page appears and displays three groupings: a list of file pool policies,
a list of template policies, and the latest scan job results.
2. In the Order column of the File Pool Policies area, select the policy that you want to
move.
3. Click either Move up or Move down until the policy is positioned where you want it in
the order.
330
Storage Pools
SmartPools settings
SmartPools settings include directory protection, global namespace acceleration, virtual
hot spare, spillover, requested protection management, and I/O optimization
management.
Setting
-
Directory
protection
Description
-
Notes
-
Increases the amount of protection for directories at a higher The option to Protect directories at one level
level than the directories and files that they contain, so that higher should be enabled.
data that is not lost can still be accessed.
When this setting is disabled, the directory that
When devices failures result in data loss (for example, three contains a file pool is protected according to your
drives or two nodes in a +2:1 policy), enabling this setting
protection-level settings, but the devices used to
ensures that intact data is still accessible.
store the directory and the file may not be the
same. There is potential to lose nodes with file
data intact but not be able to access the data
because those nodes contained the directory.
As an example, consider a cluster that has a +2
default file pool protection setting and no
additional file pool policies. OneFS directories are
always mirrored, so they are stored at 3x, which is
the mirrored equivalent of the +2 default.
This configuration can sustain a failure of two
nodes before data loss or inaccessibility. If this
setting is enabled, all directories are protected at
4x. If the cluster experiences three node failures,
although individual files may be inaccessible, the
directory tree is available and provides access to
files that are still accessible.
In addition, if another file pool policy protects
some files at a higher level, these too are
accessible in the event of a three-node failure.
Global
namespace
acceleration
Virtual hot
spare
331
Storage Pools
Setting
Description
Notes
Global spillover Specifies how to handle write operations to a node pool that
is not writable.
l
Spillover data
to
Protection
management
I/O
optimization
management
332
Subpool status
Node status
New events
Cluster size
Storage Pools
Cluster throughput
CPU usage
You also can configure real-time and historical performance to be graphed in the web
administration interface.
OneFS shortens storage pool names that are longer than 40 characters. To view the full
storage pool name, rest the mouse pointer over the shortened name to display a tooltip
of the long name.
Procedure
1. Click File System Management > SmartPools > Summary.
The SmartPools page appears and displays two groupings: the current capacity usage
and a list of storage pools.
2. In the Current Capacity Usage area, move the pointer over the usage bar-graph
measurements to view details.
3. In the Tiers & Node Pools area, expand any tiers to view all storage pool information.
333
CHAPTER 19
System jobs
System jobs
335
System jobs
To initiate any Job Engine tasks, you must have the role of SystemAdmin in the OneFS
system.
336
AutoBalance
Description
-
Exclusion Impact
Set
Policy
Restripe
Low
Priority Operation
-
Auto
System jobs
Job name
-
AutoBalanceLin
Description
-
Exclusion Impact
Set
Policy
Restripe
Low
Priority Operation
-
Auto
AVScan
Performs an antivirus
scan on all files.
None
Low
Manual
Collect
Mark
Low
Auto
Dedupe*
None
Low
Manual
None
Low
Manual
DomainMark
Low
Manual
FlexProtect
Medium
Auto
337
System jobs
Job name
-
338
FlexProtectLin
Description
-
Exclusion Impact
Set
Policy
-
Priority Operation
-
Medium
FSAnalyze
Gathers information
about the file system.
None
Low
Scheduled
IntegrityScan
Mark
Medium
Manual
MediaScan
Low
Scheduled
MultiScan
Restripe
Mark
Low
Auto
PermissionRepair
None
Low
Manual
QuotaScan*
Updates quota
None
accounting for
domains created on an
existing file tree.
Available only if you
activate a
SmartQuotas license.
Low
Auto
SetProtectPlus
Restripe
Low
Manual
None
Low
Scheduled
SmartPools*
Enforces SmartPools
file policies. Available
only if you activate a
SmartPools license.
Restripe
Low
Scheduled
SnapRevert
Reverts an entire
snapshot back to
head.
None
Low
Manual
SnapshotDelete
None
Medium
Auto
Auto
System jobs
Job name
-
TreeDelete
Description
-
Exclusion Impact
Set
Policy
-
Medium
Priority Operation
-
Manual
Job operation
OneFS includes system maintenance jobs that run to ensure that your Isilon cluster
performs at peak health. Through the Job Engine, OneFS runs a subset of these jobs
automatically, as needed, to ensure file and data integrity, check for and mitigate drive
and node failures, and optimize free space. For other jobs, for example, Dedupe, you can
use Job Engine to start them manually or schedule them to run automatically at regular
intervals.
The Job Engine runs system maintenance jobs in the background and prevents jobs
within the same classification (exclusion set) from running simultaneously. Two exclusion
sets are enforced: restripe and mark.
Restripe job types are:
u
AutoBalance
AutoBalanceLin
FlexProtect
FlexProtectLin
MediaScan
MultiScan
SetProtectPlus
SmartPools
Collect
IntegrityScan
MultiScan
Note that MultiScan is a member of both the restripe and mark exclusion sets. You
cannot change the exclusion set parameter for a job type.
The Job Engine is also sensitive to job priority, and can run up to three jobs, of any
priority, simultaneously. Job priority is denoted as 110, with 1 being the highest and 10
being the lowest. The system uses job priority when a conflict among running or queued
jobs arises. For example, if you manually start a job that has a higher priority than three
other jobs that are already running, Job Engine pauses the lowest-priority active job, runs
the new job, then restarts the older job at the point at which it was paused. Similarly, if
you start a job within the restripe exclusion set, and another restripe job is already
running, the system uses priority to determine which job should run (or remain running)
and which job should be paused (or remain paused).
Other job parameters determine whether jobs are enabled, their performance impact, and
schedule. As system administrator, you can accept the job defaults or adjust these
parameters (except for exclusion set) based on your requirements.
Job operation
339
System jobs
When a job starts, the Job Engine distributes job segmentsphases and tasksacross
the nodes of your cluster. One node acts as job coordinator and continually works with
the other nodes to load-balance the work. In this way, no one node is overburdened, and
system resources remain available for other administrator and system I/O activities not
originated from the Job Engine.
After completing a task, each node reports task status to the job coordinator. The node
acting as job coordinator saves this task status information to a checkpoint file.
Consequently, in the case of a power outage, or when paused, a job can always be
restarted from the point at which it was interrupted. This is important because some jobs
can take hours to run and can use considerable system resources.
LOW
Allowed to run
-
Resource consumption
-
Low
MEDIUM
Medium
HIGH
High
OFF_HOURS
Outside of business
hours. Business hours are
defined as 9AM to 5pm,
Monday through Friday.
OFF_HOURS is paused
during business hours.
Low
If you want to specify other than a default impact policy for a job, you can create a custom
policy with new settings.
Jobs with a low impact policy have the least impact on available CPU and disk I/O
resources. Jobs with a high impact policy have a significantly higher impact. In all cases,
however, the Job Engine uses CPU and disk throttling algorithms to ensure that tasks that
you initiate manually, and other I/O tasks not related to the Job Engine, receive a higher
priority.
Job priorities
Job priorities determine which job takes precedence when more than three jobs of
different exclusion sets attempt to run simultaneously. The Job Engine assigns a priority
value between 1 and 10 to every job, with 1 being the most important and 10 being the
least important.
The maximum number of jobs that can run simultaneously is three. If a fourth job with a
higher priority is started, either manually or through a system event, the Job Engine
pauses one of the lower-priority jobs that is currently running. The Job Engine places the
340
System jobs
paused job into a priority queue, and automatically resumes the paused job when one of
the other jobs is completed.
If two jobs of the same priority level are scheduled to run simultaneously, and two other
higher priority jobs are already running, the job that is placed into the queue first is run
first.
Start a job
By default, only some system maintenance jobs are scheduled to run automatically.
However, you can start any of the jobs manually at any time.
Procedure
1. Click Cluster Management > Job Operations > Job Summary.
2. In the Active Jobs window, select the job you want to start and click More.
Managing system jobs
341
System jobs
Pause a job
You can pause a job temporarily to free up system resources.
Procedure
1. Click Cluster Management > Job Operations > Job Summary.
2. In the Active Jobs table, click More for the job that you want to pause.
3. Click Pause Running Job in the menu that appears.
The job remains paused until you resume it.
Resume a job
You can resume a paused job.
Procedure
1. Click Cluster Management > Job Operations > Job Summary.
2. In the Active Jobs table, click More for the job that you want to pause.
3. Click Resume Running Job in the menu that appears.
Results
The job continues from the phase or task at which it was paused.
Cancel a job
If you want to free up system resources, or for any reason, you can permanently
discontinue a running, paused, or waiting job.
Procedure
1. Click Cluster Management > Job Operations > Job Summary.
2. In the Active Jobs table, click More for the job that you want to cancel.
3. Click Cancel Running Job in the menu that appears.
Update a job
You can change the priority and impact policy of a running, waiting, or paused job.
When you update a job, only the current instance of the job runs with the updated
settings. The next instance of the job returns to the default settings for that job.
Note
System jobs
343
System jobs
a. In the Add an Impact Policy Interval window, select the impact level and start and
end times from the drop-down lists.
b. Click Add Impact Policy Interval.
The Add an Impact Policy Interval window disappears, and the settings you selected
appear in the Impact Schedule table.
6. Click Create Impact Policy.
Your copy of the impact policy is saved and is listed in alphabetical order in the
Impact Policies table.
344
System jobs
Options
Description
Policy description a. In the Description field, type a new overview for the impact
policy.
b. Click Submit.
Impact schedule
The modified impact policy is saved and listed in alphabetical order in the Impact
Policies table.
345
System jobs
you to view detailed information about job load, including CPU and memory usage and
I/O operations.
346
CHAPTER 20
Networking
Networking
347
Networking
Networking overview
After you determine the topology of your network, you can set up and manage your
internal and external networks.
There are two types of networks associated with an EMC Isilon cluster:
Internal
Nodes communicate with each other using a high speed low latency InfiniBand
network. You can optionally configure a second Infiniband network as a failover for
redundancy and security.
External
Clients connect to the cluster through the external network with Ethernet. The Isilon
cluster supports standard network communication protocols, including NFS, SMB,
HTTP, and FTP. The cluster includes various external Ethernet connections, providing
flexibility for a wide variety of network configurations. External network speeds vary
by product.
You can manage both the internal and external network settings from the OneFS web
administration interface and the command-line interface.
348
Networking
Creates a default external network subnet called subnet0, with the specified
netmask, gateway, and SmartConnect service address.
Creates a default IP address pool called pool0 with the specified IP address range,
the SmartConnect zone name, and the external interface of the first node in the
cluster as the only member.
Creates a default network provisioning rule called rule0, which automatically assigns
the first external interface for all newly added nodes to pool0.
Adds pool0 to subnet0 and configures pool0 to use the virtual IP of subnet0 as its
SmartConnect service address.
Sets the global, outbound DNS settings to the domain name server list and DNS
search list, if provided.
Internal network failover
349
Networking
Once the initial external network has been established, you can configure the following
information about your external network:
u
Netmask
IP address range
Gateway
You can make modifications to the external network through the web administration
interface and the command-line interface.
IP address pools
You can partition EMC Isilon cluster nodes and external network interfaces into logical IP
address pools. IP address pools are also utilized when configuring SmartConnect zones
and IP failover support for protocols such as NFS. Multiple pools for a single subnet are
available only if you activate a SmartConnect Advanced license.
IP address pools:
u
The IP address pool of a subnet consists of one or more IP address ranges and a set of
cluster interfaces. All IP address ranges in a pool must be unique.
A default IP address pool is configured during the initial cluster setup through the
command-line configuration wizard. You can modify the default IP address pool at any
time. You can also add, remove, or modify additional IP address pools.
If you add external network subnets to your cluster through the subnet wizard, you must
specify the IP address pools that belong to the subnet.
IP address pools are allocated to external network interfaces either dynamically or
statically. The static allocation method assigns one IP address per pool interface. The IP
addresses remain assigned, regardless of that interface's status, but the method does
not guarantee that all IP addresses are assigned. The dynamic allocation method
distributes all pool IP addresses, and the IP address can be moved depending on the
interface's status and connection policy settings.
350
Networking
Advanced
If you activate a SmartConnect Advanced license, SmartConnect operates in
Advanced mode. Advanced mode enables client connection balancing based on
round robin, CPU utilization, connection counting, or network throughput. Advanced
mode supports IP failover and allows IP address pools to support multiple DNS
zones within a single subnet.
The following information describes the SmartConnect DNS client-connection balancing
policies:
Round Robin
This method selects the next available node on a rotating basis. This is the default
state (after SmartConnect is activated) if no other policy is selected.
Note
Note
SmartConnect requires that you add a new name server (NS) record to the existing
authoritative DNS zone that contains the cluster and that you delegate the SmartConnect
zone as a fully qualified domain name (FQDN).
External IP failover
You can redistribute IP addresses for external IP failover if one or more node interfaces
becomes unavailable.
SmartConnect Basic does not support IP failover. You can enable dynamic IP allocation
and IP failover in your EMC Isilon cluster if you activate a SmartConnect Advanced
license.
Dynamic IP allocation ensures that all IP addresses in the IP address pool are assigned to
member interfaces. Dynamic IP allocation allows clients to connect to any IP addresses in
the pool and receive a response. If a node or an interface becomes unavailable, OneFS
moves the IP address to other member interfaces in the IP address pool.
IP failover ensures that all of the IP addresses in the pool are assigned to an available
node. When an node interface becomes unavailable, the dynamic IP address of the node
is redistributed among the remaining available node interfaces. Subsequent client
connections are directed to the node interface that is assigned to that IP address.
External IP failover
351
Networking
If a SmartConnect Advanced license is active on the cluster, you may have enabled IP
failover when you configured your external network settings. You can also modify your
subnet settings at any time to enable IP failover for selected IP address pools.
IP failover occurs when a pool has dynamic IP address allocation set. You can further
configure IP failover for your network environment with the following options:
IP allocation method
This method ensures that all of the IP addresses in the pool are assigned to an
available node.
Rebalance policy
This policy controls how IP addresses are redistributed when the node interface
members for a given IP address pool become available after a period of
unavailability.
IP failover policy
This policy determines how to redistribute the IP addresses among remaining
members of an IP address pool when one or more members are unavailable.
NIC aggregation
Network interface card (NIC) aggregation, also known as link aggregation, is optional, and
enables you to combine the bandwidth of a node's physical network interface cards into
a single logical connection. NIC aggregation provides improved network throughput.
Note
352
Networking
Some NICs may allow aggregation of ports only on the same network card.
For LACP and FEC aggregation modes, the switch must support IEEE 802.3ad link
aggregation. Since the trunks on the network switch must also be configured, the
node must be connected with the correct ports on the switch.
VLANs
Virtual LAN (VLAN) tagging is an optional setting that enables an EMC Isilon cluster to
participate in multiple virtual networks.
You can partition a physical network into multiple broadcast domains, or virtual local
area networks (VLANs). You can enable a cluster to participate in a VLAN which allows
multiple cluster subnet support without multiple network switches; one physical switch
enables multiple virtual subnets.
VLAN tagging inserts an ID into packet headers. The switch refers to the ID to identify from
which VLAN the packet originated and to which network interface a packet should be
sent.
IPv6 support
You can configure dual stack support for IPv6.
With dual-stack support in OneFS, you can configure both IPv4 and IPv6 addresses.
However, configuring an EMC Isilon cluster to use IPv6 exclusively is not supported. When
you set up the cluster, the initial subnet must consist of IPv4 addresses.
The following table describes important distinctions between IPv4 and IPv6.
IPv4
-
32-bit addresses
Subnet mask
IPv6
-
128-bit addresses
Prefix length
Modify the IP address ranges of the internal network and the int-b/failover network
VLANs
353
Networking
You can configure the int-b/failover network to provide backup in the event of an int-a
network failure. Configuration involves specifying a valid netmask and IP address range
for the failover network.
3. In the IP Ranges area, you can add, delete, or migrate your IP address ranges. Ideally
the new range is contiguous with the previous one. For example, if your current IP
address range is 192.168.160.60 - 192.168.160.162, the new range should start with
192.168.160.163.
4. Click Submit.
For the changes in netmask value to take effect, you must reboot the cluster.
Procedure
1. Click Cluster Configuration > Network Configuration.
2. In the Internal Network Settings area, select the network that you want to configure
the netmask for.
l
354
To immediately reboot the cluster, click Yes. When the cluster finishes rebooting,
the login page appears.
Networking
Click No to return to the Edit Internal Network page without changing the settings
or rebooting the cluster.
355
Networking
Adding a subnet
You can add and configure an external subnet.
Adding a subnet to the external network encompasses these tasks:
Procedure
1. Configuring subnet settings.
2. Adding an IP address to a new subnet.
3. Optional: Configuring SmartConnect settings for a new subnet.
4. Selecting interface members for a new subnet.
For an IPv4 subnet, click IPv4 in the IP Family list. In the Netmask field, type a
dotted decimal octet (x.x.x.x) that represents the subnet mask.
For an IPv6 subnet, click IPv6 in the IP family list. In the Prefix length field, type an
integer (ranging from 1 to 128) that represents the network prefix length.
6. In the MTU list, type or select the size of the maximum transmission units the cluster
uses in network communication. Any numerical value is allowed, but must be
compatible with your network and the configuration of all devices in the network path.
Common settings are 1500 (standard frames) and 9000 (jumbo frames).
Although OneFS supports both 1500 MTU and 9000 MTU, using a larger frame size for
network traffic permits more efficient communication on the external network
between clients and cluster nodes.
Note
To benefit from using jumbo frames, all devices in the network path must be
configured to use jumbo frames.
356
Networking
7. In the Gateway address field, type the IP address of the gateway server device through
which the cluster communicates with systems outside of the subnet.
8. In the Gateway priority field, type an integer for the priority of the subnet gateway for
nodes assigned to more than one subnet.
You can configure only one default gateway per node, but each subnet can be
assigned a gateway. When a node belongs to more than one subnet, this option
enables you to define the preferred default gateway. A value of 1 represents the
highest priority, and 10 represents the lowest priority.
9. If you plan to use SmartConnect for connection balancing, in the SmartConnect
service IP field, type the IP address that will receive all incoming DNS requests for
each IP address pool according to the client connection policy. You must have at least
one subnet configured with a SmartConnect service IP in order to use connection
balancing.
10.Optional: In the Advancedsection, you can enable VLAN tagging if you want to enable
the cluster to participate in virtual networks.
Note
If your cluster is running SmartConnect Basic for connection balancing, you can configure
only one IP address pool per subnet. If you activate a SmartConnect Advanced license,
you can configure unlimited IP address pools per subnet.
Procedure
1. In the Step 2 of 4 IP Address Pool Settings dialog box, type a unique Name for the
IP address pool. The name can be up to 32 alphanumeric characters long and can
include underscores or hyphens, but no spaces or other punctuation.
2. Type a Description for the IP address pool. The description can contain up to 128
characters.
Adding a subnet
357
Networking
3. In the Access zone list, click to select an access zone for the pool. OneFS includes a
default system access zone.
4. In the IP range (low-high) area, click New.
OneFS adds an IP address range with default Low IP and High IP values.
5. Click to select the default Low IP value. Replace the default value with the starting IP
address of the subnet's IP address pool.
6. Click to select the default High IP value. Replace the default value with the ending IP
address of the subnet's IP address pool.
7. Optional: Add IP address ranges to the IP address pool by repeating steps 3 through 6
as needed.
8. Click Next.
The Step 3 of 4 SmartConnect Settings dialog box appears.
After you finish
The next step in the process of adding a new subnet is configuring SmartConnect
settings, which is optional. If you do not wish to configure SmartConnect settings, the
next step is adding network interface members to the new subnet.
SmartConnect requires that you add a new name server (NS) record to the existing
authoritative DNS zone that contains the cluster and that you delegate the SmartConnect
zone as a fully qualified domain name (FQDN).
Procedure
1. In the Step 3 of 4 SmartConnect Settings dialog box, type a Zone name for the
SmartConnect zone that this IP address pool represents. The zone name must be
unique among the pools served by the SmartConnect service subnet specified in Step
3 below.
2. In the Connection policy list, select the type of connection balancing policy set by the
IP address pool of this subnet. The connection balancing policy determines how
SmartConnect distributes incoming DNS requests across the members of an IP
address pool.
358
Networking
Options
Description
Round Robin
Connection Count
Network Throughput Sets the overall average throughput volume on each available
node to optimize the cluster usage.
CPU Usage
3. In the SmartConnect service subnet list, select the name of the external network
subnet whose SmartConnect service will answer DNS requests on behalf of the IP
address pool. A pool can have only one SmartConnect service answering DNS
requests. If this option is left blank, the IP address pool the subnet belongs to is
excluded when SmartConnect answers incoming DNS requests for the cluster.
Note
If you have activated a SmartConnect Advanced license, complete the following steps
for the options in the SmartConnect Advanced section of this wizard page.
4. In the IP allocation method list, select the method by which IP addresses are assigned
to the member interfaces for this IP address pool:
Options
Description
Static
Dynamic Select this IP allocation method to ensure that all IP addresses in the IP
address pool are assigned to member interfaces, which allows clients to
connect to any IP addresses in the pool and be guaranteed a response. If
a node or an interface becomes unavailable, their IP addresses are
automatically moved to other available member interfaces in the pool.
If you select the dynamic IP allocation method, you can specify the SmartConnect
Rebalance policy and the IP failover policy in the next two steps.
5. Select the type of SmartConnect Rebalance policy to redistribute IP addresses. IP
address redistribution occurs when node interface members in an IP address pool
become available. These options can only be selected if the IP allocation method is
set to Dynamic.
Adding a subnet
359
Networking
Options
Description
Automatic
Automatically redistributes IP addresses. The automatic
Failback (default) rebalance is triggered by a change to one of the following items.
Manual Failback
Cluster membership.
Description
Round Robin
Connection Count
Network Throughput Sets the overall average throughput volume on each available
node to optimize the cluster usage.
CPU Usage
7. Click Next to store the changes that you made to this wizard page.
The Step 4 of 4 IP Address Pool members dialog box appears.
After you finish
The next step in the process of adding a new subnet is adding network interface
members.
Networking
In the case of aggregated links, choose the aggregation mode that corresponds to the
switch settings from the Aggregation mode drop-down.
Note
Modifying an external network subnet that is in use can disable access to the cluster and
the web administration interface. OneFS displays a warning if deleting a subnet will
terminate communication between the cluster and the web administration interface.
Procedure
1. Click Cluster Management > Network Configuration.
2. In the External Network Settings area, click the name of the subnet you want to
modify.
3. In the Settings area, click Edit.
4. Modify the Basic subnet settings as needed.
Options
Description
Description
Netmask
The subnet mask for the network interface. This field appears
only for IPv4 subnets.
MTU
Gateway address
Gateway priority
The priority of the subnet's gateway for nodes that are assigned
to more than one subnet. Only one default gateway can be
configured on each Isilon node, but each subnet can have its
own gateway. If a node belongs to more than one subnet, this
option enables you to define the preferred default gateway. A
Managing external network subnets
361
Networking
Options
Description
value of 1 is the highest priority, with 10 being the lowest
priority.
SmartConnect
service IP
Description
VLAN tagging
VLAN ID
Hardware load
balancing IPs
You can enter the IP address for a hardware load balancing switch
that uses Direct Server Return (DSR).
6. Click Submit.
Networking
3. To verify that the static route was created, run the following command: isi networks
ls pools -v.
OK
3. To ensure that the static route was created, run the following command: isi
networks ls pools -v.
363
Networking
6. Click Submit.
Networking
Configure IP failover
You can configure IP failover to reassign an IP address from an unavailable node to a
functional node, which enables clients to continue communicating with the cluster, even
after a node becomes unavailable.
Procedure
1. Click Cluster Management > Network Configuration
2. In the External Network Settings area, click the name of the subnet for which you
want to set up IP failover.
3. Expand the area of the pool you want to modify and click Edit in the SmartConnect
Settings area.
4. Optional: In the Zone name field, enter a name for the zone, using no more than 128
characters.
5. In the Connection Policy list, select a balancing policy:
Options
Description
Round Robin
Connection Count
Network Throughput Uses the overall average throughput volume on each available
node to optimize the cluster usage.
CPU Usage
6. If you purchased a license for SmartConnect Advanced, you will also have access to
the following lists:
IP allocation method
This setting determines how IP addresses are assigned to clients. Select either
Dynamic or Static.
Rebalance Policy
This setting defines the client redirection policy for when a node becomes
unavailable. The IP allocation list must be set to Dynamic in order for rebalance
policy options to be selected.
Managing IP address pools
365
Networking
IP failover policy
This setting defines the client redirection policy when an IP address becomes
unavailable.
366
Networking
367
Networking
4. In the Available interfaces table, click the aggregated interface for the node, which is
indicated by a listing of AGGREGATION in the Type column.
For example, if you want to aggregate the network interface card for Node 2 of the
cluster, click the interface named ext-agg, Node 2 under Available interfaces, and
then click the right-arrow button to move the aggregated interface to the Interfaces in
current pool table.
5. From the Aggregation mode drop-down, select the appropriate aggregation mode that
corresponds to the network switch settings.
Note
Consult your network switch documentation for supported NIC aggregation modes.
OneFS supports the following NIC aggregation modes:
Options
Description
Link Aggregation
Control Protocol
(LACP)
Legacy Fast
EtherChannel (FEC)
mode
Etherchannel (FEC)
Active / Passive
Failover
Round-Robin Tx
6. Click Submit.
368
LNI numbering corresponds to the physical positioning of the NIC ports as found on
the back of the node. LNI mappings are numbered from left to right.
Aggregated LNIs are listed in the order in which they are aggregated at the time they
are created.
Networking
LNI
-
ext-1
NIC
-
em0
Aggregated LNI
-
lagg0
fec0
ext-2
em1
ext-1
em2
lagg0
fec0
ext-2
em3
lagg1
fec1
ext-3
em0
lagg2
fec2
ext-4
em1
ext-1
em0
lagg0
fec0
ext-2
em1
lagg1
fec1
10gige-1 cxgb0
10gige-1 cxgb1
369
Networking
Setting DNS resolver options may change how OneFS performs DNS lookups. Do not
set DNS resolver options unless directed to do so by Isilon Technical Support.
6. Click Submit.
Networking
You might have already configured SmartConnect while setting up an external network
subnet using the Subnet wizard. However, you can configure or modify connection
balancing settings at any time as your networking requirements change.
Procedure
1. Click Cluster Management > Network Configuration.
2. In the External Network Settings area, click the link for the subnet that you want to
configure for connection balancing.
3. In the Settings area, verify that the SmartConnect service IP was configured.
If the SmartConnect service IP field reads Not set, click Edit, and then specify the IP
address that DNS requests are directed to.
4. In the SmartConnect settings area , click Edit.
5. In the Zone name field, type a name for the SmartConnect zone that this IP address
pool represents. The zone name must be unique among the pools served by the
SmartConnect service subnet that is specified in Step 7 below.
6. In the Connection policy drop-down list, select the type of connection balancing
policy that is configured for the IP address pool for this zone. The policy determines
how SmartConnect distributes incoming DNS requests across the members of an IP
address pool.
Note
Round robin is the only connection policy available if you have not activated a
SmartConnect Advanced license.
Options
Description
Round Robin
Connection Count
Network Throughput Sets the overall average throughput volume on each available
node to optimize the cluster usage.
CPU Usage
7. In the SmartConnect service subnet list, select the name of the external network
subnet whose SmartConnect service answers DNS requests on behalf of the IP
address pool. A pool can have only one SmartConnect service answering DNS
requests. If this option is left blank, the IP address pool that the SmartConnect service
belongs to is excluded when SmartConnect answers incoming DNS requests for the
cluster.
If you have activated a SmartConnect Advanced license, complete the following steps
in the SmartConnect Advanced area.
8. In the IP allocation method list, select the method by which IP addresses are assigned
to the member interfaces for this IP address pool.
371
Networking
372
Options
Description
Any
Apply the provisioning rule to all types of Isilon nodes that join
the cluster.
Storage-i
Accelerator-i
Storage
Accelerator
Networking
Options
Description
Description
ext-1
ext-2
ext-3
ext-4
ext-agg
ext-agg-2
ext-agg-3
ext-agg-4
10gige-1
10gige-2
373
Networking
374
CHAPTER 21
Hadoop
Hadoop
375
Hadoop
Hadoop overview
Hadoop is a flexible, open-source framework for large-scale distributed computation.
The OneFS file system can be configured for native support of the Hadoop Distributed File
System (HDFS) protocol, enabling your cluster to participate in a Hadoop system.
HDFS integration requires you to activate a separate license. To obtain additional
information or to enable HDFS support for your EMC Isilon cluster, contact your EMC Isilon
sales representative.
376
CDH 3
CDH 4
CDH 4.2
Greenplum HD 1.1
Greenplum HD 1.2
Pivotal HD 1.0.1
HAWQ 1.1.0.1
Hadoop
Managing HDFS
To keep the HDFS service performing efficiently on an Isilon cluster, you will need to be
familiar with the user and system configuration options available as part of an HDFS
implementation.
You can manage an HDFS implementation by using the following methods:
u
Hadoop client machines are configured directly through their Hadoop installation
directory.
A secure shell (SSH) connection to a node in the Isilon cluster is used to configure the
HDFS service.
You can combine multiple options with a single isi hdfs command. For command
usage and syntax, run the isi hdfs -h command.
2. To set the default logging level for the Hadoop daemon across the cluster, run the isi
hdfs command with the --log-level option.
Valid values are listed below, in descending order from the highest to the lowest
logging level. The default value is NOTICE. The values are case-sensitive.
l
system database.
l
ERR: Errors.
NOTICE: Conditions that are not error conditions, but may need special handling.
377
Hadoop
DEBUG: Messages that contain information typically of use only when debugging a
program.
For example, the following command sets the log level to WARNING:
isi hdfs --log-level=WARNING
3. To set the path on the cluster to present as the HDFS root directory, run the isi
hdfs command with the --root-path option.
Valid values include any directory path beginning at /ifs, which is the default HDFS
root directory.
For example, the following command sets the root path to /ifs/hadoop:
isi hdfs --root-path=/ifs/hadoop
4. To set the Hadoop block size, run the isi hdfs command with the --block-size
option.
Valid values are 4KB to 1GB. The default value is 64MB.
For example, the following command sets the block size to 32 MB:
isi hdfs --block-size=32MB
5. To tune the number of worker threads that HDFS uses, run the isi hdfs command
with the --num-threads option.
Valid values are 1 to 256 or auto, which is calculated as twice the number of cores.
The default value is auto.
For example, the following command specifies 8 worker threads:
isi hdfs --num-threads=8
6. To allocate IP addresses from an IP address pool, run the isi hdfs racks
modify command.
The following command allocates IP addresses from a pool named pool1, which is in
the subnet0 subnet.
isi hdfs racks modify --name=/default-rack --add-ippool=subnet0:pool1
Note
378
Hadoop
The cluster is running properly with HDFS in simple security authentication mode.
System clocks on the Isilon nodes and the Hadoop clients are synchronized with a
formal time source such as Active Directory or NTP.
You know how to configure Linux and Unix systems to work with Kerberos.
You know how to manage an EMC Isilon cluster through the command-line interface.
You must perform the following tasks in sequence to configure HDFS authentication
through Kerberos.
1. Configure the realm, hostnames, and domain for HDFS.
2. Configure OneFS for Kerberos.
3. Configure the krb5.conf file.
4. Configure cluster settings for HDFS.
5. Modify Hadoop configuration files for Kerberos authentication.
For more information about authenticating HDFS connections through Kerberos, see the
white paper EMC Isilon Best Practices for Hadoop Data Storage.
379
Hadoop
You have already configured a Kerberos system with a resolvable hostname for the
KDC and a resolvable hostname for the KDC admin server.
You know how to configure client computers to work with Kerberos authentication.
You have established an SSH connection to a node in the cluster and are logged in as
the root user.
2. Then run the following command to set the default realm, where <REALM> is the name
that you specified for the realm name:
isi auth krb5 modify default --default-realm=<REALM>
4. Run the following command to verify that you added the realm correctly:
kadmin -p <admin-principal-name>
kadmin: quit
You have already configured a Kerberos system with a resolvable hostname for the
KDC and a resolvable hostname for the KDC admin server.
You know how to configure client computers to work with Kerberos authentication.
You have created a SmartConnect zone for your Isilon cluster. If you have not, see
Managing external client connections with SmartConnect.
You configure the cluster to use the principal that matches the name of the SmartConnect
zone that you created. This principal must be added to the KDC and exported to the
keytab file, which is then copied to the nodes in the cluster.
Procedure
1. Run the following command to set the HDFS principal to be used and return the
principal that you must add to your KDC, where <SmartConnect-zone-name> is the name
380
Hadoop
of the SmartConnect zone that you created for your cluster. You must be logged on as
the root user to run this command:
isi hdfs krb5 --kerb-instance=<SmartConnect-zone-name>
2. Run the following commands on the cluster with the kadmin utility to create a
principal on the KDC, where <admin-principal-name> is the value for the admin
credentials for your KDC, and the principal that is being created is the one returned
from the previous command:
kadmin -p <admin-principal-name>
add_principal -randkey hdfs/<SmartConnect-zone-name>
3. In the kadmin utility, run the following command to add the principal to a keytab file
and export the keytab file with a unique file name:
ktadd -k <keytab /path/filename> hdfs/<SmartConnect-zone-name>
kadmin: quit
4. In the OneFS command-line interface, securely copy the keytab file to all the nodes in
your cluster. You can script this copying task if you have several nodes. The following
command is just one example of a script that copies the keytab files to every node:
for ip in `isi_nodes %{internal}`;
do scp <keytab /path/filename> $ip:/etc/;
done
The cluster is running properly with HDFS in simple security authentication mode.
You know how to configure Linux and Unix systems to work with Kerberos and Active
Directory.
You know how to manage an EMC Isilon cluster through the command-line interface.
You must perform the following tasks in sequence to configure HDFS authentication
through Active Directory Kerberos.
Configuring HDFS authentication with Active Directory Kerberos
381
Hadoop
After you configure Kerberos authentication, you cannot use a local user on a Hadoop
compute client to access HDFS data on an EMC Isilon cluster; you must use an account in
Active Directory.
Procedure
1. With standard Active Directory tools, create HDFS user accounts, similar to the
following patterns for the three service principal names, where <jt-username>@<DOMAIN> and <tt-user-name>@<DOMAIN> are placeholders for JobTracker and
TaskTracker user names:
hdfs@<DOMAIN>
<jt-user-name>@<DOMAIN>
<tt-user-name>@<DOMAIN>
2. Establish a secure connection such as SCP and distribute the keytab files to the
Hadoop compute clients that need them.
382
Hadoop
Note
The Kerberos keytab file contains an encrypted, local copy of the host's key, which if
compromised may potentially allow unrestricted access to the host computer. It is
crucial to protect the keytab file with file-access permissions.
You know how to configure client computers to work with Kerberos authentication.
You have established an SSH connection to a node in the cluster and are logged in as
the root user.
Procedure
1. Run the following commands to join the cluster to the Active Directory domain, where
<DOMAIN> is a placeholder for your domain name.
Note
You can skip this command if you are already joined to a domain:
isi auth ads create <DOMAIN> administrator
isi auth ads modify --provider-name=<DOMAIN> --assume-defaultdomain=true
krb5.conf
mapred-site.xml
hdfs-site.xml
core-site.xml
You must restart the Hadoop daemons on the compute clients to apply the changes.
383
Hadoop
Hadoop
<value>hdfs/<hostname>@<REALM | DOMAIN></value>
</property>
<property>
<name>dfs.datanode.keytab.file</name>
<value>/etc/krb5.keytab</value>
</property>
<property>
<name>dfs.datanode.kerberos.principal</name>
<value>hdfs/<hostname>@<REALM | DOMAIN></value>
</property>
</configuration>
385
Hadoop
To run Hadoop jobs or distcp, you must make sure that the principals that the Hadoop
daemons are using, the value of the TaskTrackers and JobTrackers Kerberos principal in
mapred-site.xml, map to users on the Isilon cluster and can be resolved on the
cluster by using either OneFS local users or users from LDAP or Active Directory.
Procedure
1. Open the file mapred-site.xml with a text editor.
2. Add the principal for the JobTracker and the location of its keytab file to mapredsite.xml, using the following example for configuring two compute clients with a
service principal name (SPN):
<?xml version="1.0"?>
<!-- mapred-site.xml -->
<configuration>
<property>
<name>mapreduce.jobtracker.kerberos.principal</name>
<value><jt-user-name>/_HOST@<REALM | DOMAIN></value>
</property>
<property>
<name>mapreduce.jobtracker.keytab.file</name>
<value>/etc/<jt-user-name>.keytab</value>
</property>
Next, add the principal for the TaskTracker and the location of its keytab file, using the
following example:
<property>
<name>mapreduce.tasktracker.kerberos.principal</name>
<value><tt-user-name>/_HOST@<REALM | DOMAIN></value>
</property>
<property>
<name>mapreduce.tasktracker.keytab.file</name>
<value>/etc/<tt-user-name>.keytab</value>
</property>
</configuration>
You have already configured a Kerberos system with a resolvable hostname for the
KDC and a resolvable hostname for the KDC admin server.
Test the connection from a compute client and then run a sample MapReduce job to
verify your configuration.
386
Hadoop
Procedure
1. On a compute client, run the following commands to validate the connection to the
cluster:
su hdfs
kinit hdfs@<REALM | DOMAIN>
$HADOOP_HOME/bin/hadoop fs -ls /
The following command example runs on an Isilon node named kdc-demo-1. The name of
the MIT Kerberos 5 KDC and admin server is york.east.example.com.
kdc-demo-1# isi auth krb5 add realm \
--realm=EAST.EXAMPLE.COM --kdc=york.east.example.com \
--admin-server=york.east.example.com
kdc-demo-1# isi auth krb5 modify default \
--default-realm=EAST.EXAMPLE.COM
kdc-demo-1# isi auth krb5 write
kdc-demo-1# kadmin -p root/admin@EAST.EXAMPLE.COM
Authenticating as principal root/admin@EAST.EXAMPLE.COM with password.
Password for root/admin@EAST.EXAMPLE.COM:
kadmin: quit
kdc-demo-1# ping wai-kerb-sc
PING wai-kerb-sc.east.example.com (192.0.2.11): 56 data bytes
64 bytes from 192.0.2.11: icmp_seq=0 ttl=64 time=0.561 ms
kdc-demo-1# isi hdfs krb5 --kerb-instance=wai-kerb-sc
Add this principal to your KDC: \
hdfs/wai-kerb-sc.east.example.com@<YOUR-REALM.COM>
kdc-demo-1# kadmin -p root/admin
Authenticating as principal root/admin with password.
Password for root/admin@EAST.EXAMPLE.COM:
kadmin: add_principal -randkey hdfs/wai-kerb-sc.east.example.com
WARNING: no policy specified for
hdfs/wai-kerb-sc.east.example.com@EAST.EXAMPLE.COM;
defaulting to no policy
Principal "hdfs/wai-kerb-sc.east.example.com@EAST.EXAMPLE.COM"
created.
kadmin: ktadd -k /ifs/hdfs.keytab hdfs/wai-kerb-sc.east.example.com
Entry for principal hdfs/wai-kerb-sc.east.example.com with kvno 3,
Sample commands for configuring MIT Kerberos authentication over HDFS
387
Hadoop
encryption type AES-256 CTS mode with 96-bit SHA-1 HMAC added to
keytab WRFILE:/ifs/hdfs.keytab.
Entry for principal hdfs/wai-kerb-sc.east.example.com with kvno 3,
encryption type ArcFour with HMAC/md5 added to keytab
WRFILE:/ifs/hdfs.keytab.
Entry for principal hdfs/wai-kerb-sc.east.example.com with kvno 3,
encryption type Triple DES cbc mode with HMAC/sha1 added to keytab
WRFILE:/ifs/hdfs.keytab.
Entry for principal hdfs/wai-kerb-sc.east.example.com with kvno 3,
encryption type DES cbc mode with CRC-32 added to keytab
WRFILE:/ifs/hdfs.keytab.
kdc-demo-1# for ip in `isi_nodes %{internal}`;
do scp /ifs/hdfs.keytab
$ip:/etc/;done
Password:
hdfs.keytab 100% 666 0.7KB/s 0.7KB/s 00:00
Max throughput: 0.7KB/s
kdc-demo-1# kinit -k -t /etc/hdfs.keytab \
hdfs/wai-kerb-sc.east.example.com
kdc-demo-1# klist
Ticket cache: FILE:/tmp/krb5cc_0
Default principal:
hdfs/wai-kerb-sc.east.example.com@EAST.EXAMPLE.COM
Valid starting Expires Service principal
01/28/14 15:15:34 01/29/14 01:15:34
krbtgt/EAST.EXAMPLE.COM@EAST.EXAMPLE.COM
renew until 01/29/14 15:13:46 kdc-demo-1# kdestroy
kdc-demo-1# isi hdfs krb5 --keytab=/etc/hdfs.keytab
388
Check all the configuration parameters, including the location and validity of the
keytab file.
Check user and group accounts for permissions. Make sure that there are no
duplicate accounts across systems, such as a local hdfs account on OneFS and an
hdfs account in Active Directory.
Make sure that the system clocks on the Isilon nodes and the Hadoop clients are
synchronized with a formal time source such as Active Directory or NTP. The Kerberos
standard requires that system clocks be no more than 5 minutes apart.
Check to confirm that the service principal name of a Hadoop service, such as
TaskTracker, is not mapped to more than one object in Active Directory.
CHAPTER 22
Antivirus
Antivirus
389
Antivirus
Antivirus overview
You can scan the files you store on an Isilon cluster for computer viruses and other
security threats by integrating with third-party scanning services through the Internet
Content Adaptation Protocol (ICAP). OneFS sends files through ICAP to a server running
third-party antivirus scanning software. These servers are referred to as ICAP servers.
ICAP servers scan files for viruses.
After an ICAP server scans a file, it informs OneFS of whether the file is a threat. If a threat
is detected, OneFS informs system administrators by creating an event, displaying near
real-time summary information, and documenting the threat in an antivirus scan report.
You can configure OneFS to request that ICAP servers attempt to repair infected files. You
can also configure OneFS to protect users against potentially dangerous files by
truncating or quarantining infected files.
Before OneFS sends a file to be scanned, it ensures that the scan is not redundant. If a
file has already been scanned and has not been modified, OneFS will not send the file to
be scanned unless the virus database on the ICAP server has been updated since the last
scan.
Note
Antivirus scanning is available only if all nodes in the cluster are connected to the
external network.
On-access scanning
You can configure OneFS to send files to be scanned before they are opened, after they
are closed, or both. Sending files to be scanned after they are closed is faster but less
secure. Sending files to be scanned before they are opened is slower but more secure.
If OneFS is configured to ensure that files are scanned after they are closed, when a user
creates or modifies a file on the cluster, OneFS queues the file to be scanned. OneFS then
sends the file to an ICAP server to be scanned when convenient. In this configuration,
users can always access files without any delay. However, it is possible that after a user
modifies or creates a file, a second user might access the file before the file is scanned. If
a virus was introduced to the file from the first user, the second user will be able to
access the infected file. Also, if an ICAP server is unable to scan a file, the file will still be
accessible to users.
If OneFS ensures that files are scanned before they are opened, when a user attempts to
download a file from the cluster, OneFS first sends the file to an ICAP server to be
scanned. The file is not sent to the user until the scan is complete. Scanning files before
they are opened is more secure than scanning files after they are closed, because users
can access only scanned files. However, scanning files before they are opened requires
users to wait for files to be scanned. You can also configure OneFS to deny access to files
that cannot be scanned by an ICAP server, which can increase the delay. For example, if
no ICAP servers are available, users will not be able to access any files until the ICAP
servers become available again.
If you configure OneFS to ensure that files are scanned before they are opened, it is
recommended that you also configure OneFS to ensure that files are scanned after they
are closed. Scanning files as they are both opened and closed will not necessarily
improve security, but it will usually improve data availability when compared to scanning
files only when they are opened. If a user wants to access a file, the file may have already
390
Antivirus
been scanned after the file was last modified, and will not need to be scanned again if
the ICAP server database has not been updated since the last scan.
The name and IP address of the user that triggered the scan.
This information is not included in reports triggered by antivirus scan policies.
391
Antivirus
ICAP servers
The number of ICAP servers that are required to support an Isilon cluster depends on how
virus scanning is configured, the amount of data a cluster processes, and the processing
power of the ICAP servers.
If you intend to scan files exclusively through antivirus scan policies, it is recommended
that you have a minimum of two ICAP servers per cluster. If you intend to scan files on
access, it is recommended that you have at least one ICAP server for each node in the
cluster.
If you configure more than one ICAP server for a cluster, it is important to ensure that the
processing power of each ICAP server is relatively equal. OneFS distributes files to the
ICAP servers on a rotating basis, regardless of the processing power of the ICAP servers. If
one server is significantly more powerful than another, OneFS does not send more files to
the more powerful server.
McAfee VirusScan Enterprise 8.7 and later with VirusScan Enterprise for Storage 1.0
and later.
Antivirus
You can configure OneFS and ICAP servers to react in one of the following ways when
threats are detected:
Repair or quarantine
Attempts to repair infected files. If an ICAP server fails to repair a file, OneFS
quarantines the file. If the ICAP server repairs the file successfully, OneFS sends the
file to the user. Repair or quarantine can be useful if you want to protect users from
accessing infected files while retaining all data on a cluster.
Repair or truncate
Attempts to repair infected files. If an ICAP server fails to repair a file, OneFS
truncates the file. If the ICAP server repairs the file successfully, OneFS sends the file
to the user. Repair or truncate can be useful if you do not care about retaining all
data on your cluster, and you want to free storage space. However, data in infected
files will be lost.
Alert only
Only generates an event for each infected file. It is recommended that you do not
apply this setting.
Repair only
Attempts to repair infected files. Afterwards, OneFS sends the files to the user,
whether or not the ICAP server repaired the files successfully. It is recommended that
you do not apply this setting. If you only attempt to repair files, users will still be
able to access infected files that cannot be repaired.
Quarantine
Quarantines all infected files. It is recommended that you do not apply this setting. If
you quarantine files without attempting to repair them, you might deny access to
infected files that could have been repaired.
Truncate
Truncates all infected files. It is recommended that you do not apply this setting. If
you truncate files without attempting to repair them, you might delete data
unnecessarily.
Click Only scan files smaller than the maximum file size and specify a maximum
file size.
3. In the Filename restrictions area, specify whether to exclude files from being scanned
based on file names and extensions.
Configuring global antivirus settings
393
Antivirus
Click Scan all files except those with the following extensions or filenames.
4. Optional: If you chose to exclude files based on file names and extensions, specify
which files will be selected.
a. In the Extensions area, click Edit list, and specify extensions.
b. In the Filenames area, click Edit list, and specify filenames.
You can specify the following wild cards:
-
Wildcard Description
-
[ ]
5. Click Submit.
To require that all files be scanned before they are opened by a user, select Scan
files when they are opened, and then specify whether you want to allow access to
files that cannot be scanned.
To scan files after they are closed, select Scan files when they are closed.
3. In the Directories to be scanned area, specify the directories that you want to apply
on-access settings to.
394
Antivirus
If no directories are specified, on-access scanning settings are applied to all files. If
you specify a directory, only files from the specified directories will be scanned as
they are accessed.
4. Click Submit.
395
Antivirus
396
Antivirus
6. In the Run policy area, specify whether you want to run the policy according to a
schedule or manually.
Scheduled policies can also be run manually at any time.
Options
Description
Click Manually
a. Click Scheduled.
b. In the Interval area, specify on what days you want
the policy to run.
c. In the Frequency area, specify how often you want
the policy to run on the specified days.
7. Click Submit.
397
Antivirus
Scan a file
You can manually scan an individual file for viruses. This procedure is available only
through the command-line interface (CLI).
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
2. Run the isi avscan manual command.
For example, the following command scans /ifs/data/virus_file:
isi avscan manual /ifs/data/virus_file
398
Antivirus
Rescan a file
You can rescan the file for viruses if, for example, you believe that a file is no longer a
threat.
Procedure
1. Click Data Protection > Antivirus > Detected Threats.
2. In the Detected Threats table, in the row of a file, click Rescan.
399
Antivirus
View threats
You can view files that have been identified as threats by an ICAP server.
Procedure
1. Click Data Protection > Antivirus > Detected Threats.
2. In the Detected Threats table, view potentially infected files.
400
Antivirus
401
CHAPTER 23
iSCSI
iSCSI overview.....................................................................................................404
iSCSI targets and LUNs........................................................................................ 404
iSNS client service...............................................................................................405
Access control for iSCSI targets........................................................................... 405
iSCSI considerations and limitations................................................................... 406
Supported SCSI mode pages............................................................................... 406
Supported iSCSI initiators................................................................................... 407
Configuring the iSCSI and iSNS services.............................................................. 407
Managing iSCSI targets....................................................................................... 409
Configuring iSCSI initiator access control............................................................ 411
Creating iSCSI LUNs.............................................................................................415
Managing iSCSI LUNs.......................................................................................... 418
iSCSI
403
iSCSI
iSCSI overview
The Isilon iSCSI module enables customers to provide block storage for Microsoft
Windows, Linux, and VMware systems over an IP network. To access the iSCSI module,
you must activate a separate license.
Note
The iSCSI module is supported by exception only and is limited to certain use cases. It is
recommended that you find other solutions for your block storage requirements.
You can create and manage iSCSI targets on a cluster. The targets are available as SCSI
block devices on which you can store structured and unstructured data. iSCSI targets
contain one or more logical units, each uniquely identified by a logical unit number (LUN).
You can format and connect to targets, such as physical disk devices, on the local file
system.
You can configure each target to limit access to a list of initiators. You can also require
initiators to authenticate with a target by using the Challenge-Handshake Authentication
Protocol (CHAP).
The iSCSI module includes the following features:
u
Isilon FlexProtect
LUN cloning
404
Normal
A normal LUN is the default LUN type for clone and imported LUNs, and the only type
available for newly created LUNs. Normal LUNs are either writeable or read-only.
Snapshot
Clone
A clone LUN is a copy of a normal, snapshot, or clone LUN. Clone LUNs are
implemented using overlay and mask files in conjunction with a snapshot. Clone
iSCSI
LUNs require little time and disk space to create, and the LUN is fully writeable. You
can create clone LUNs by cloning or importing existing LUNs.
CHAP authentication
You can authenticate initiator connections to iSCSI targets with the Challenge-Handshake
Authentication Protocol (CHAP).
You can restrict initiator access to a target by enabling CHAP authentication and then
adding user:secret pairs to the target's CHAP secrets list. If you enable CHAP
authentication, initiators are required to provide a valid user:secret pair to authenticate
their connections to the target. CHAP authentication is disabled by default.
Note
The Isilon iSCSI module does not support mutual CHAP authentication.
405
iSCSI
Multipath I/O (MPIO) is recommended only for iSCSI workflows that have primarily
read-only operations. The node must invalidate the data cache on all other nodes
during file-write operations and performance decreases in proportion to the number
of write operations. If all MPIO sessions are connected to the same node,
performance should not decrease.
The Isilon iSCSI module supports the importing of normal LUNs only. Importing
snapshot LUNs and clone LUNs is not supported. You cannot back up and then
restore a snapshot or clone LUN, or replicate snapshot or clone LUNs to another
cluster. It is recommended that you deploy a backup application to back up iSCSI
LUNs on the iSCSI client, as the backup application ensures that the LUN is in a
consistent state at the time of backup.
406
08h
00h
3Fh
00h
0Ah
00h
00h
iSCSI
* For the caching mode page, OneFS supports the write cache enable (WCE) parameter
only.
** OneFS supports querying this mode page through the Mode Sense command, but
does not support changing the fields of this page through the Mode Select command.
iSCSI Initiator
-
Microsoft Windows 2003 (32-bit and 64-bit) Microsoft iSCSI Initiator 2.08 or later (Certified)
Microsoft Windows 2008 (32-bit and 64-bit) Microsoft iSCSI Initiator (Certified)
Microsoft Windows 2008 R2 (64-bit only)
All of the current iSCSI sessions will be terminated for all the nodes in the cluster.
Initiators cannot establish new sessions until the iSCSI service is re-enabled.
Procedure
1. Click File System Management > iSCSI > Settings.
2. In the iSCSI Service area, set the service state that you want:
l
407
iSCSI
iSNS server address: Type the IP address of the iSNS server with which you want to
register iSCSI target information.
iSNS server port: Type the iSNS server port number. The default port number is
3205.
If the service is disabled, you can enable it by clicking Enable. Enabling the service
allows OneFS to register information about iSCSI targets.
If the service is enabled, you can disable it by clicking Disable. Disabling the
service prevents OneFS from registering information about iSCSI targets.
To view historical iSCSI throughput data, you must obtain the EMC Isilon InsightIQ virtual
appliance, which requires you to activate a separate license. For more information,
contact your EMC Isilon representative.
Procedure
1. Click File System Management > iSCSI > Summary.
2. Review the current throughput data and current session information.
408
The Current Throughput area displays a chart that illustrates overall inbound and
outbound throughput across all iSCSI sessions during the past hour, measured in
kilobits per second (Kbps). This chart automatically updates every 15 seconds.
The Current Sessions area displays information about each current connection
between an initiator and a target, including the client and target IP addresses;
node, target, and LUN; operations per second; and the inbound, outbound, and
total throughput in bits per second. You can view details about a target by clicking
the target name.
iSCSI
This directory is used only if no other directory is specified during LUN creation or if a
LUN is not created. The directory must be in the /ifs directory tree. The full path to
the directory is required, and wildcard characters are not supported.
6. Add one or more SmartConnect pools for the target to connect with. This setting
overrides any global default SmartConnect pools that are configured for iSCSI targets.
a. For the SmartConnect pool(s) setting, click Edit list.
b. Move pools between the Available Pools and Selected Pools lists by clicking a
pool and then clicking the right or left arrow. To remove all selected pools at once,
click clear.
c. Click OK.
7. Click Submit.
8. Optional: In the Initiator Access Control area, enable and configure the settings for
initiator access control.
a. Click Enable to restrict target access to initiators that are added to the initiator
access control settings.
b. Click Add initiator.
409
iSCSI
c. In the Initiator name field, type the name of the initiator that you want to allow to
access this target, or click Browse to select from a list of initiators. An initiator
name must begin with an iqn. prefix.
d. Click OK.
Note
To continue adding initiators, click OK and add another. When you are finished adding
initiators, click OK.
9. Optional: In the CHAP Authentication area, enable and configure ChallengeHandshake Authentication Protocol (CHAP) settings.
Note
If CHAP authentication is enabled and the CHAP secrets list is empty, no initiators can
access the target.
a. Click Enable to require initiators to authenticate with the target.
b. Click Add username.
c. In the Username field, type the name that the initiator will use to authenticate with
the target. You can specify an initiator's iSCSI qualified name (IQN) as the
username. Depending on whether you specify an IQN, valid usernames differ in the
following ways:
If you specify an IQN as the username, the Username value must begin with an
iqn. prefix. The characters that are allowed after the iqn. prefix are
alphanumeric characters, periods (.), hyphens (-), and colons (:).
All other usernames can use alphanumeric characters, periods (.), hyphens (-),
and underscores (_).
Note
410
iSCSI
Note
l
Changing the default path does not affect existing logical units.
4. Click Submit.
Target Details: Displays the target name, IQN, description, default path, capacity,
and SmartConnect pool settings. The name and IQN cannot be modified.
Logical Units: Displays any logical units that are contained in the target. You can
add or import a logical unit, or manage existing logical units. You can also select
the columns to display or hide.
Allowed Initiators: Displays the target's initiator access control status, and lists
the names of any initiators that are allowed to access the target when access
control is enabled.
CHAP Authentication: Displays the target's CHAP authentication status, and lists
all user:secret pairs for the target.
411
iSCSI
If access control is disabled, click Enable to restrict target access to initiators that
you add to the initiator access list.
Note
If you disable access control and the initiator access list is empty, no initiators are
able to connect to the target.
l
If access control is enabled, click Disable to allow all initiators access to the
target.
Note
iSCSI
If you remove all of the allowed initiators for a target and access control is enabled,
the target will deny new connections until you disable access control.
Removing an allowed initiator for a target does not affect the initiator's access to
other targets.
Procedure
1. Click File System Management > iSCSI > Targets & Logical Units.
2. In the Targets area, under Actions, click Edit for the target that you want to modify.
3. In the Initiator Access Control area, under Actions, click Delete for the initiator that
you want to remove from the access list.
4. In the confirmation dialog box, click Yes.
413
iSCSI
If you specify an IQN as the username, the Username value must begin with an
iqn. prefix. The characters that are allowed after the iqn. prefix are alphanumeric
characters, periods (.), hyphens (-), and colons (:).
All other usernames can use alphanumeric characters, periods (.), hyphens (-), and
underscores (_).
Note
If you delete all of a target's CHAP secrets and CHAP authentication is enabled, no
initiators are able to access the target until you disable CHAP authentication.
Procedure
1. Click File System Management > iSCSI > Targets & Logical Units.
2. In the Targets area, under Actions, click Edit for the target that you want to delete a
CHAP user:secret pair for.
3. In the CHAP Authentication area, under Actions, click Delete for the CHAP user:secret
pair that you want to delete.
4. In the confirmation dialog box, click Yes.
414
iSCSI
Procedure
1. Click File System Management > iSCSI > Targets & Logical Units.
2. In the Targets area, under Actions, click Edit for the target whose CHAP authentication
state you want to modify.
3. In the CHAP Authentication area, configure the initiator's CHAP authentication state.
l
If CHAP authentication is enabled and the CHAP secrets list is empty, no initiator is
able to access the target.
l
When you create a logical unit, you must assign it to an existing iSCSI target. Each target
can contain one or more logical units.
Procedure
1. Click File System Management > iSCSI > Targets & Logical Units.
2. In the Logical Units area, click Add logical unit.
3. In the Add Logical Unit area, in theDescription field, type a descriptive comment for
the logical unit.
4. From the Target list, select the target that will contain the logical unit.
Creating iSCSI LUNs
415
iSCSI
To assign the next available number to the logical unit, click Automatic. This is the
default setting.
To manually assign a number to the logical unit, click Manual and then, in the
Number field, type an integer value. The value must be within the range 0-255
and must not be assigned to another logical unit within the target.
By default, the LUN number forms part of the directory name that is created for storing
the LUN data.
6. To manually specify the path where the LUN directory is created, in the Path field, type
the full path of the directory, beginning with /ifs, or click Browse to select the
directory.
The directory must be in the /ifs directory tree. You must specify the full path to the
directory, and wildcard characters are not allowed. The default path is /ifs/iscsi/
ISCSI.LUN.<TargetName>.<LUNnumber>, where <TargetName> is the Target
value and <LUNnumber> is the LUN number.
7. In the Size field, specify the LUN capacity by typing an integer value and then
selecting a unit of measure from the list (MB, GB, or TB).
The minimum LUN size is 1 MB. The maximum LUN size is determined by the OneFS
file system. After you create a LUN, you can increase its size, but you cannot decrease
it.
8. Select one of the Provisioning options.
l
To specify that blocks are unallocated until they are written, click Thin provision.
To immediately allocate all the blocks, click Pre-allocate space. This is the default
setting.
Note
Allocation of all the blocks for a large LUN can take hours or even days.
9. Select one of the LUN access options.
l
To make the LUN accessible, click Online. This is the default setting.
To allow iSCSI initiators to write to the LUN, click Read-Write. This is the default
setting.
11.Under Protection Settings, from the Disk pool list, select the disk pool to contain the
logical unit.
12.From the SSD strategy list, select to specify a strategy to use if solid-state drives
(SSDs) are available.
416
Metadata read acceleration (Recommended): Writes metadata and all user data on
hard disk drives (HDDs) and additionally creates a mirror backup of the metadata
on an SSD. Depending on the global namespace acceleration setting, the SSD
mirror may be an extra mirror in addition to the number required to satisfy the
protection level.
iSCSI
Data on SSDs (Requires most SSD space): Similar to metadata acceleration, but
also writes one copy of the file's user data (if mirrored) or all of the data (if not
mirrored) on SSDs. Regardless of whether global namespace acceleration is
enabled, any SSD blocks reside on the file's target pool if there is room. This SSD
strategy does not create additional mirrors beyond the normal protection level.
Avoid SSDs (Reduces performance): Never uses SSDs; writes all associate file data
and metadata to HDDs only.
13.From the Protection level list, select a protection policy for the logical unit. Select Use
iSCSI default (2x), which is the recommended setting for best performance, or one of
the mirrored options, such as 2x to 8x.
14.Select one of the Write Cache options.
l
To prevent write caching for files that contain LUN data, click Disabled. This is the
recommended setting for LUNs.
To allow write caching for files that store LUN data, click Enable.
Note
The Write Cache option controls whether file writes are sent to the coalescer or the
endurant cache. With Write Cache disabled, which is the default and recommended
setting, all file writes are sent to the endurant cache. The endurant cache is a
committed data guarantee. If Write Cache is enabled, all file writes are sent to the
coalescer. Write caching can improve performance, but can lead to data loss if a node
loses power or crashes while uncommitted data is in the write cache.
15.Select one of the Data access pattern options.
l
To select a random access pattern, click Random. This is the recommended setting
for LUNs.
16.Click Submit.
The settings for the clone vary according to the source LUN type.
5. Click Submit.
417
iSCSI
Clone LUN
type
-
Normal
Normal
Result
-
A snapshot of the source LUN is created. The clone LUN is then created by copying the LUN data
from the snapshot. After completing the copy, the snapshot is deleted. The copy process may take
several hours to complete for large LUNs if the source LUN has a pre-allocated provisioning policy.
The copy process may also take several minutes for thinly provisioned LUNs that are significantly
used.
Normal
Snapshot
A snapshot of the source LUN is created. The clone LUN is configured to reference the data from
the snapshot. The snapshot is deleted when the clone is deleted.
Normal
Clone
A snapshot of the source LUN is created. The system then creates a clone LUN that references data
from the snapshot.
Snapshot
Normal
The clone LUN is created by copying the LUN data from the snapshot. The copy process may take
several minutes to complete for large LUNs if the source LUN has a pre-allocated provisioning
policy. The copy process may also take several minutes for thinly provisioned LUNs that are heavily
used.
Snapshot
Snapshot
The clone LUN is configured to reference the data from the same snapshot that the source LUN
references. The underlying snapshot is not deleted when a LUN is deleted unless the LUN being
deleted is the last LUN referencing the snapshot.
Snapshot
Clone
The clone LUN is configured to reference the data from the same snapshot that the source LUN
references. The underlying snapshot is not deleted when a LUN is deleted unless the LUN being
deleted is the only LUN referencing the snapshot.
Clone
Normal
A snapshot of the source LUN is created. The clone LUN is then created by copying the LUN data
from the snapshot. After completing the copy, the snapshot is deleted. The copy process may take
several minutes to complete for large LUNs if the source LUN has a pre-allocated provisioning
policy. The copy process may also take several minutes for thinly provisioned LUNs that are heavily
used
Clone
Snapshot
Not allowed.
Clone
Clone
A clone of the clone LUN is created. The clone LUN is configured to reference data from the
snapshot.
418
Modify a LUN
Delete a LUN
iSCSI
Import a LUN
To assign the next available number to the logical unit, click Automatic. This is the
default setting.
To manually assign a number to the logical unit, click Manual and then, in the
Number box, type an integer value. The value must be within the range 0-255 and
must not be assigned to another logical unit.
5. To configure the path where the LUN directory is created, in the To path box, type the
full path of the directory, or click Browse to select the directory. If a path is not
specified, the LUN directory is unchanged from the original directory where that LUN
was created.
6. Click Submit.
419
iSCSI
To assign the next available number to the logical unit, click Automatic. This is the
default setting.
To manually assign a number to the logical unit, click Manual, and then in the
Number field, type an integer value. The value must be within the range 0-255 and
must not be assigned to another logical unit.
To make the LUN accessible, click Online. This is the default setting.
To allow iSCSI initiators to write to the LUN, click Read-Write. This is the default
setting.
To allow write caching for files storing LUN data, click Enabled.
To prevent write caching for files storing LUN data, click Disabled.
10.Click Submit.
420
LUN: Displays the numerical identifier of the logical unit. You can modify the LUN
value by using the move operation.
iSCSI
Target: Displays the name of the iSCSI target that contains the logical unit. You can
modify the target by using the move operation.
Description: Displays an optional description for the logical unit. You can modify
the description by clicking Edit LUN.
Type: Displays the LUN type (normal, clone, or snapshot). You cannot modify this
setting.
Size: Displays the LUN capacity. You can increase the size of normal or snapshot
LUNs by clicking Edit LUN, but you cannot decrease the size. You cannot modify
the size of snapshot LUNs.
Status: Displays the connection status (online or offline) and write access
permissions (read-only or read-write) of the LUN. You can modify write-access
settings for normal or clone LUNs by clicking Edit LUN. You cannot modify writeaccess settings for snapshot LUNs.
Path: Displays the path to the directory where the LUN files are stored. You can
change the path for normal or snapshot LUNs by using the move operation. You
cannot modify the path for snapshot LUNs.
Disk pool: Displays the disk pool of the LUN. You can modify the disk pool by
clicking Edit LUN.
Protection level: Displays the mirroring level (such as 2x, 3x, 4x, and so on) or
FlexProtect protection policy for the LUN. You can modify the protection policy for
normal or clone LUNs by clicking Edit LUN. You cannot modify these settings for
snapshot LUNs.
Write Cache: Displays whether SmartCache is enabled or disabled. You can change
this setting for normal or clone LUNs by clicking Edit LUN. You cannot modify these
settings for snapshot LUNs.
Data access pattern: Displays the access pattern setting (Random, Concurrency, or
Streaming) for the LUN. You can change the access pattern for normal or clone
LUNs by clicking Edit LUN. You cannot modify these settings for snapshot LUNs.
SCSI name: Displays the iSCSI qualified name (IQN) of the LUN. You cannot modify
this setting.
EUI: Displays the extended unique identifier (EUI), which uniquely identifies the
LUN. You cannot modify this setting.
NAA: Displays the LUN's T11 Network Address Authority (NAA) namespace. You
cannot modify this setting.
Serial number: Displays the serial number of the LUN. You cannot modify this
setting.
421
CHAPTER 24
VMware integration
VMware integration
423
VMware integration
VAAI
OneFS uses VMware vSphere API for Array Integration (VAAI) to support offloading
specific virtual machine storage and management operations from VMware ESXi
hypervisors to an Isilon cluster.
VAAI support enables you to accelerate the process of creating virtual machines and
virtual disks. For OneFS to interact with your vSphere environment through VAAI, your
VMware environment must include ESXi 5.0 or later hypervisors.
If you enable VAAI capabilities for an Isilon cluster, when you clone a virtual machine
residing on the cluster through VMware, OneFS clones the files related to that virtual
machine. For more information on file clones, see Clones.
The iSCSI module is supported by exception only and is limited to certain use cases. It is
recommended that you find other solutions for your block storage requirements.
OneFS supports the following VAAI primitives for block storage:
u
Full Copy
Block Zeroing
Note
OneFS does not support the thin provisioning block reclaim mechanism.
424
VMware integration
VASA
OneFS communicates with VMware vSphere through VMware vSphere API for Storage
Awareness (VASA).
VASA support enables you to view information about Isilon clusters through vSphere,
including Isilon-specific alarms in vCenter. VASA support also enables you to integrate
with VMware profile driven storage by providing storage capabilities for Isilon clusters in
vCenter. For OneFS to communicate with vSphere through VASA, your VMware
environment must include ESXi 5.0 or later hypervisors.
Description
-
Thin-provisioned LUN There is not enough available space on the cluster to allocate space for
capacity exceeded
writing data to thinly provisioned LUNs. If this condition persists, you will
not be able to write to the virtual machine on this cluster. To resolve this
issue, you must free storage space on the cluster.
The Isilon I-Series and X-Series nodes contain Solid State Drives (SSDs). If a cluster
is composed of i-Series, X-Series , or S-Series nodes, but does not contain SSDs, the
cluster is recognized as a capacity cluster.
425
VMware integration
Capacity
The Isilon cluster is composed of Isilon X-Series nodes that do not contain SSDs. The
cluster is configured for a balance between performance and capacity.
Hybrid
The Isilon cluster is composed of nodes associated with two or more storage
capabilities. For example, if the cluster contained both Isilon S-Series and NL-Series
nodes, the storage capability of the cluster is displayed as Hybrid.
Enable VASA
You must enable an Isilon cluster to communicate with VMware vSphere API for Storage
Awareness (VASA) by enabling the VASA daemon.
Procedure
1. Open a secure shell (SSH) connection to any node in the cluster and log in.
2. Enable VASA by running the following command:
isi services isi_vasa_d enable
Record the location of where you saved the certificate. You will need this file path
when adding the vendor provider in vCenter.
VMware integration
2. Fill out the following fields in the Add Vendor Provider window:
Name
Type a name for this VASA provider. Specify as any string. For example, type EMC
Isilon Systems.
URL
Type https://<IPAddress>:8081/vasaprovider, where <IPAddress> is the IP
address of a node in the Isilon cluster.
Login
Type root.
Password
Type the password of the root user.
Certificate location
Type the file path of the vendor provider certificate for this cluster.
3. Select the Use Vendor Provider Certificate box.
4. Click OK.
3. Disable or enable the VASA daemon by running one of the following commands:
l
427
CHAPTER 25
File System Explorer
429
The File System Explorer displays up to 1000 files in a directory. If more than 1000 files
exist within a directory, the files are displayed without additional information, such as file
size and last modified date.
The contents of the selected directory are displayed in the right pane. You can view
the contents of another directory by clicking the directory in the Directories pane.
Create a directory
You can create a directory under /ifs through the File System Explorer.
Procedure
1. Navigate to File System Management > File System Explorer.
2. In the Directories pane, specify where you want to create the directory.
3. Click Add Directory.
4. In the New Directory Properties dialog box, in the Directory name field, type a name
for the directory.
5. From the User list, select the owner of the directory.
6. From the Group list, select the group for the directory.
7. From the Permissions table, specify the basic permissions for the directory.
8. Click Submit.
430
431
SSD
The SSD strategy that will be used for user data and metadata if solid-state drives
(SSDs) are available. The following SSD strategies are available:
Metadata acceleration
OneFS creates a mirror backup of file metadata on an SSD and writes the rest of
the metadata plus all user data to hard disk drives (HDDs). Depending on the
global namespace acceleration setting, the SSD mirror might be an extra mirror
in addition to the number required to satisfy the protection level.
Avoid SSDs
OneFS does not write data or metadata to SSDs. OneFS writes all data and
metadata to HDDs only.
Data on SSDs
Similar to metadata acceleration, OneFS creates a mirror backup of file
metadata on an SSD and writes the rest of the metadata plus all user data to
hard disk drives. However, OneFS also writes one copy of the file user data (if
mirrored) or all of the data (if not mirrored) to SSDs. All SSD blocks reside on the
file target pool if there is adequate space available, regardless of whether
global namespace acceleration is enabled. OneFS does not create additional
mirrors beyond the normal protection level.
Actual protection
The FlexProtect or data-mirroring requested protection for this file or directory. If
SmartPools is licensed and enabled on the cluster, the default requested protection
for files and directories is inherited from the specified disk pool.
I/O Optimization Settings
Settings Management
Specifies whether I/O Optimization Settings are managed manually or by
SmartPools. If you modify either or both I/O optimization settings, this property
automatically refreshes to Manually managed. If you specify Managed by
SmartPools, the I/O optimization settings values will automatically refresh to match
the SmartPools specifications the next time the SmartPools job is run.
SmartCache
Specifies whether write caching with SmartCache is enabled for this file or directory.
432
433