Distributed File Systems-2
Distributed File Systems-2
Transparent local access — Data to be accessed as if it’s local to the user for high
performance.
Location independence — No need for users to know where file data physically
resides.
Scale-out capabilities — The ability to scale out massively by adding more machines.
DFS systems can scale to exceedingly large clusters with thousands of servers.
Fault tolerance — A need for your system to continue operating properly even if some
of its servers or disks fail. A fault-tolerant DFS is able to handle such failures by
spreading data across multiple machines.
Lustre
GlusterFS
Ceph
Distribution: First, a DFS distributes datasets across multiple clusters or nodes. Each
node provides its own computing power, which enables a DFS to process the datasets
in parallel.
Replication: A DFS will also replicate datasets onto different clusters by copying the
same pieces of information into multiple clusters. This helps the distributed file system to
achieve fault tolerance—to recover the data in case of a node or cluster failure—as well
as high concurrency, which enables the same piece of data to be processed at the
same time.
To enable enterprises to take back control of their data at scale, Cohesity has built a
completely new file system: SpanFS. SpanFS is designed to effectively consolidate and
manage all secondary data, including backups, files, objects, dev/test, and analytics
data, on a web-scale, multicloud platform that spans from core to edge to cloud.
With Cohesity SpanFS, you can consolidate data silos across locations by uniquely
exposing industry-standard, globally distributed NFS, SMB, and S3 protocols on a single
platform.
These are among the top benefits of SpanFS:
Unlimited scalability — Start with as little as three nodes and grow limitlessly on-
premises or in the cloud and pay-as-you-grow.
Automated global indexing — Perform powerful global actionable wildcard searches
for any virtual machine (VM), file, or object.
Guaranteed data resiliency — Maintain strict consistency across nodes within a
cluster to ensure data resiliency at scale.
Dedupe across workloads and clusters — Reduce your data footprint with global
variable-length dedupe across workloads and protocols.
Cloud-ready — Use the Cohesity Helios multicloud data platform to eliminate
dependency on bolt-on cloud gateways.
Multiprotocol access — Seamlessly read and write to the same data volume with
simultaneous multiprotocol access for NFS, SMB, and S3.