Symantec Topic Takeover on Storage
Data management tools from searchStorage.com
Video -When used in conjunction with processes like multi-copy mirroring and erasure coding, object storage is a reliable form of storage for rapidly growing amounts of data.
Tip -Be aware: Copy data management is easily misunderstood. For example, it is not a true backup and could impact performance. Find out how best to get around these issues.
E-Chapter -Because big data can scale to petabytes of capacity, organizations are looking to manage it in ways that are easier and less expensive than traditional scale-out NAS. Object storage and software-defined storage are frequently mentioned as big data tools. Both can add intelligence required for analyzing data and take advantage of low-cost disk storage.
An object storage system handles files differently than a traditional file system. Servers use unique identifiers to find objects, which use metadata in a far more detailed way than file systems do. The unique identifiers mean objects can be geographically dispersed because they can be retrieved without the storage system knowing their physical location. That makes objects a good choice for large data stores or data stored in a cloud.
Software-defined storage has many forms and use cases, but it applies to big data when used to pool and manage data across off-the-shelf commodity hardware. Because the management and analytics happen in software appliances, the hardware can be cheap, deep disk without bells and whistles.
Perhaps the most well known option available is the Apache Hadoop Distributed File System (HDFS), which is a Java-based file system designed to be used in Hadoop clusters. HDFS currently scales to 200 petabytes and can support single Hadoop clusters of 4,000 nodes. It offers storage performance on a large scale and at a low cost, which is atypical of most enterprise arrays that cannot perform all three tasks simultaneously.
In this chapter of "Tools to Tackle Big Data Troubles," we look at some core HDFS features, three HDFS commercial distributions and other Hadoop storage-related tools and their related applications.
Tiered storage from searchStorage.com
News -Quantum goes after video workloads with Xcellis, a converged system that combines Quantum StorNext, metadata, data management, NAS and cloud gateways into a single system.
News -New FileFly software for Caringo Swarm lets Windows Storage Server and NetApp filers migrate data to Caringo object storage systems.
News -Data scientists who study topics such as disease and climate change require robust storage systems to hold petabytes of crucial research.
Solid state storage technology from searchSolidStateStorage.com
E-Handbook -Arrays loaded with flash drives are the speed kings of storage arrays. They can also be more expensive than traditional systems with hard disk drives (HDDs), so IT managers need to assess their application needs to make the best decision among all-flash array vendors.
Cheap HDDs give traditional arrays an advantage in price per GB, but systems equipped only with more costly flash drives can provide a significant edge in price per IOPS. Hybrid arrays combining HDDs and solid-state drives (SSDs) are yet another option to consider when balancing the price-performance equation.
Points of comparison for IT organizations weighing the various options from all-flash array (AFA) vendors include IOPS, latency, throughput, raw and usable capacity, flash type, networking options, architecture type (scale-up vs. scale-out), and supported storage features, such as data deduplication and compression, thin provisioning, snapshots, replication and encryption.
The original use case for AFAs was typically to accelerate the performance of a niche application with high I/O requirements. AFAs have since become more popular for primary storage, running multiple application workloads on a single flash array now that the products offer the capacity, management and storage capabilities to put them on par with HDD-based systems.
The use of denser, less expensive flash, such as multi-level cell and triple-level cell 3D NAND, and data reduction technologies are giving rise to claims from AFA vendors that their products can match or beat the price of high-end storage arrays equipped with the fastest spinning disks. So IT organizations need to compare the features and capabilities of many of the leading AFAs against specific criteria to select the right AFA to meet their technical and business needs.
E-Handbook -All-flash arrays are a hot technology, but not everybody needs flash for all of their storage. Hybrid flash arrays can strike a balance between using flash for performance while keeping spinning disk drives to lower the price for less frequently accessed data. Flash storage offers blazing speed but at a high cost per gigabyte.
At the other end of the spectrum, multi-terabyte hard disk drives (HDDs) are more economical, but they do not supply the raw IOPS per drive that some applications need. Hybrid flash arrays combining HDDs and a thin slice of flash storage can provide a performance boost and reduce latency while keeping costs in check. Although the difference between HDD prices and flash costs has narrowed considerably, many organizations still don't have the budget to deploy hundreds of terabytes of solid-state storage. Despite differences in architectures, the vendors generally agree on some hybrid vs. all-flash guidelines. If sub-millisecond latency or guaranteed quality of service (QoS) is required, then an all-flash array or a hybrid flash array that can deliver near all-flash performance is the way to go. But with variable and unpredictable workloads, hybrid flash arrays can often serve the need at a lower $/GB.
Candidates for hybrid flash arrays include collaboration, email and any applications where data lifecycle issues mean that not all data requires immediate access.
News -Tegile Systems and Pure Storage are pushing more cost-effective flash upgrade programs that are customized for the IT utility model for all-flash arrays.