Author Archives: mesabigroup

IBM SpectrumAI with NVIDIA DGX Reference Architecture: A Solid Foundation for AI Data Infrastructures

IBM Storage and NVIDIA have teamed up to enhance artificial intelligence (AI) project development and streamline the AI data pipeline. This approach — IBM SpectrumAI with NVIDIA DGX Reference Architecture — provides data scientists and other AI project team members with a solid framework that can help in AI deployments and ends with design based on IBM and NVIDIA system and software products.

The companies’ partnership is important not only because the field of AI is growing very rapidly, but because major AI projects can be a real challenge to any organization. IBM Storage in combination with NVIDIA and its joint channel partners offers skills, resources and products to enable organizations to overcome whatever challenges they might face for their AI workloads.

The AI Revolution

Information technology (IT) always seems to be in the throes of a major revolution. AI is one such revolution and despite all that is going on, AI is still in its infancy. Many years hence, AI may still not even be at the knee of the curve of a decades-long exponential growth. Every day it seems that there is a new or expanded practical use of AI technology — such as self-driving cars, a huge number of customer sentiment and sensor-based analysis examples, threat analysis, and image interpretation. Almost all organizations should be able to benefit from AI technology, now or in the future. And infrastructure vendors are thrilled by the prospect since AI projects often demand seemingly inexhaustible compute and storage resources.

From reference architecture to a converged infrastructure solution

AI projects are data-driven in contrast to the process orientation of online transaction processing systems (OLTP). An AI data pipeline consists of ingest, classification and analyzing/training phases that require considerable development time and thought, so an AI reference architecture can substantially aid the efforts of project teams. In general, reference architectures are increasing in popularity as they provide a frame of reference for a particular domain. Reference architectures are available for specific industries and processes, such as banking, telecommunications, and manufacturing and supply chains.

These play an important role, but so can vendor-supplied reference architectures, such as the IBM SpectrumAI with NVIDIA DGX reference architecture. Vendor-specific reference architectures lead AI project teams down a path to purchasing products that implement an AI infrastructure solution. This is not a problem if AI project teams understand up front what they are getting into and are comfortable with the vendors.

The roles of IBM and NVIDIA in the IBM SpectrumAI with NVIDIA DGX Reference Architecture

Most, if not all, organizations should be comfortable with IBM and NVIDIA, two of the giants in the AI industry. Of course, IBM Watson is familiar to many, but the company has strengths and expertise in non-Watson-related AI activities. NVIDIA notably invented the GPU (Graphical Processing Unit), which has become a chief computing element in AI (such as in NVIDIA DGX servers), where it serves as an accelerator for the highly dense parallel processing engine AI projects typically demand. This is now complemented on the storage side by IBM Spectrum Scale, which, at its software-storage-sysem-based heart, has a long-proven and well-accepted parallel file system that enables close integration with DGX servers.

The net result — a powerful combination of IBM and NVIDIA for AI workloads — which encompasses all the necessary computing, storage and networking hardware that is accompanied by all the required supporting software in a single physical rack put together by IBM’s and NVIDIA’s channel partners.

The system consists of NVIDIA DGX-1 servers with Tesla V100 Tensor Core GPUs for computing. IBM supplies the storage solution with ESS (Enterprise Storage Servers) GS4S (All-Flash, non-NVMe) storage systems for immediate use, but moving to NVMe flash arrays in mid-2019 according to IBM (and that should be sufficient time as typical large AI projects have a significant gestation period). Mellanox IB (InfiniBand) Networking provides the necessary connectivity between the servers and storage elements.

But don’t forget the software.  The NVIDIA DGX software stack is specifically designed to deliver GPU-accelerated training performance, and that includes the new RAPIDS framework whose purpose is to accelerate data science workflow. At the heart of the IBM software-defined-storage (SDS) for files is IBM Spectrum Scale v5, which was specifically architected for the high-performance demand of modern AI workloads.

Now, NVIDIA’s arrangement with IBM is not an exclusive one — DDN Storage, NetApp and Pure Storage also work with the company on AI-related solutions — so how does IBM differentiate itself from these strong competitors? IBM claims that it has a performance advantage, stating that it will have a 1.5x NVMe advantage against competitors.  Additionally, IBM Spectrum Scale has extensive use in AI workloads already, including two AI reference architectures with IBM Power servers, and vast experience in the HPC-like needs of AI use cases.

IBM SpectrumAI with NVIDIA DGX will be sold only through selected channel partners supported by both companies. This makes a great deal of sense as major AI projects require a level of planning and design knowledge, along with collaboration and coordination skills, that only selected channel partners can bring to the table.

Mesabi musings

If you have not already done so, the time may be right to hop on the AI bandwagon. If you agree, looking into vendor-sponsored reference architectures, such as the one featured with IBM SpectrumAI with NVIDIA DGX, might be a good starting point. Just be sure that you realize that these vendors will eventually propose an AI deployment involving their products.

Still, you are not planning such efforts just for the fun of it, so eventually a converged infrastructure solution could provide an ideal way forward. IBM and NVIDIA are both leaders in their respective parts of the AI domain and their new IBM SpectrumAI with NVIDIA DGX offering makes a strong case for the companies.

IBM Driving Storage Revolutions

Business storage continues to be driven by two revolutions: one is storage systems–based and the other software-based. The former is focused on NVMe (nonvolatile memory express) networking technology that is accelerating the adoption of all-flash storage systems. In the latter case, software-driven innovation has become a driving force among virtually all major storage vendors.

 

One vendor that is making notable progress in both areas is IBM. On the systems/network side, i.e. NVMe-oF (NVM over Fabrics), IBM now supports Fibre Channel in addition to InfiniBand. Additionally, the company’s new Storwize V7000 Gen 3 has been architected for NVMe at the storage array level as well, joining the FlashSystem 9100 family (announced in July) with NVMe inside of the storage array. On the storage software side, IBM has just introduced Spectrum Discover as a new product in its IBM Spectrum Storage portfolio. Let’s examine these additions in a little more detail.

 

IBM continues to push the NVMe revolution

 

NVMe has two basic functions. NVMe-oF is the network side of the house and improves the performance of moving data between a host and a storage array. IBM initially enabled NVMe-oF for storage networks that use InfiniBand interconnects but now supports NVMe-oF with storage networks that use Fibre Channel (FC) to improve application performance and data access. This functionality runs in conjunction with the company’s Spectrum Virtualize through a straightforward, non-disruptive software upgrade. FC NVMe uses existing 16 Gb FC adapters and supports SVC (Model SV1), FlashSystem 9100, FlashSystem V9000 (Model AC3), and Storwize V7000F/V7000 Gen 2+ and Gen 3, and VersaStack that uses those storage arrays. This is likely to be important for users of those systems, as many of them likely have a FC SAN (storage area network).

 

IBM also continues to push NVMe at the storage device-level. Recall that the FlashSystem 9100, IBM’s enterprise-class entree in the virtual storage infrastructure space managed by Spectrum Virtualize was the first IBM storage system to offer NVMe at the device level. (See https://mesabigroup.com/ibm-flashsystem-9100-the-importance-of-nvme-based-storage-in-a-data-driven-multi-cloud-world/ for more detail.) Now, the new Storwize V7000 Gen 3–also managed by Spectrum Virtualize–offers the same NVMe end-to-end capability. That includes the use of the same version of IBM’s well-accepted FlashCore Modules that the FlashSystem 9100 pioneered.

 

Although the Storwize V7000 Gen 3 is technically not an all-flash solution (as users have the option to have some HDDs, such as for supporting non-performance-sensitive data), it can be configured as an all-flash system, and with the notable growth of all-flash arrays over the past few years, Mesabi Group expects a high percentage of them to be all-flash configurations. Since only flash (not hard disks) can benefit from NVMe technology at the device level, IT can maximize its use of a Storwize V7000 Gen 3 by having as much of its storage as feasible reside on flash storage modules (the new Storwize V7000 supports both IBM’s FlashCore technology as well as industry standard NVMe SSDs) instead of HDDs. If they do, Gen 3 offers up to a 2.7x throughput performance improvement over Gen 2+ as a key benefit.

 

IBM Spectrum Discover drives additional value from oceans of unstructured data

 

IT must get the most out of its investment in its physical architecture. For storage management purposes, that includes how storage arrays work in conjunction with the servers that demand services through a storage network. IBM’s storage management software, Storage Insights, is an AI-based tool that is offered through IBM Cloud to help users better manage their storage environments. For example, the latest version diagnoses storage network “gridlock” issues often referred to as “slow drain”. That gridlock occurs when a storage system attempts to send data to a server faster than the server can accept it; this is not a good thing! IBM storage technicians (who can monitor systems on behalf of clients who authorize it) are notified by Storage Insights of the problem as it is identified by AI technology. The technicians then review the situation and work with the client to resolve it.

 

Now, while Storage Insights deals with the physical side of storage as a storage management tool, recently announced IBM Spectrum Discover is an in-house data management software tool that targets the voluminous, and ever-rapidly growing amount of data, such as that created for Internet of Things (IoT), AI, and Big data analytic applications. Spectrum Discover works with file data managed by IBM Spectrum Scale or object data managed by IBM Cloud Object Storage, and enables users to get more out of their data for analytical, governance and storage investment purposes (IBM will also support Dell/EMC’s Isilon offerings in 2019).

 

How does it accomplish this? On the analytical side, getting to useful and actionable insights that would not be discovered otherwise within a data ocean of unstructured data rapidly is facilitated by such things as its ability to orchestrate machine learning and MapReduce processes. On the governance side, mitigating business risks by ensuring that data is compliant with governance policies and speeding up investigation into potentially fraudulent activities obviously may be of great value. On the investment side, the ability to facilitate the movement of “colder” (i.e., less frequently-accessed data suitable, say, for archiving) data to cheaper storage and to weed out and destroy unnecessary redundant data is financially advantageous.

 

The heart of Spectrum Discover’s power revolves around its metadata management and related processes. Any search and discover tool needs good data about data (i.e. metadata) to succeed. Spectrum Discover uses both automatically-generated system metadata at the time of data creation and custom metadata tagging that adds extra intelligence that is needed at analysis time. All that leads to automatic cataloging with the creation of an index where large quantities of data can be searched extremely rapidly for discovery purposes, thus reducing data scientist and storage admin preparation time and costs associated with that.

 

Although for different purposes (and not totally similar technologies) as an analogy, think of the search and discover capabilities of a public Internet browser for speed and flexibility for publicly-available data in contrast to the private data that Spectrum Discover deals with. Accompanying search and discover functions are a number of features and capabilities that greatly facilitate the use of the tool, including policy-driven workflows, a drill down dashboard, and an Action Agent that manages data movement and facilitates content inspection.

 

In essence, IBM Spectrum Discover is designed to significantly simplify and speed the data and storage processes required for analytics and AI processes. That should provide notable benefits for enterprises that aim to maximize the effectiveness and value of their advanced analytics investments.

 

Mesabi musings

You would think that storage innovations would show signs of slowing down after all these years, but the opposite seems to be true. In fact, IBM continues to be at the forefront of storage progress.

 

As illustrations of its continuing leadership, IBM has introduced the new NVMe-enabled Storwize V7000 Gen 3 on the systems side of storage, Spectrum Discover on the software side as a data management tool, and enhanced Spectrum Insights as a storage management tool.

 

Overall, IBM customers should be pleased with the progress IBM is making with NVMe technology, a fundamental storage systems underpinning technology on the hardware side, while Storage Discover, on the software side, continues the push toward extracting additional value from up to oceans of unstructured data.

IBM Continues to Deliver New Multicloud Storage Solutions

The ever-increasing vast quantities of data that need to be stored, distributed, and managed cost-effectively with security and reliability are looking to the multicloud for a solution. IBM recognizes this and is delivering new multicloud storage solutions in response to that need.

Overview of Multicloud and IBM Storage’s Multicloud Solutions

The movement to the multicloud represents a dramatic shift in how the majority of enterprise class IT organizations will restructure their information infrastructure now and in the coming years. A multicloud represents at least two cloud environments — say a private and public cloud — but more typically describes the use of multiple public clouds typically in conjunction with an on-premises or private cloud. Enterprises want to be able to move applications and data swiftly and easily from place to place to best handle workload requirements (such as performance and availability) while at the same time generating the best possible cost-efficiencies. That demand for agility and flexibility comes with the challenge of providing the necessary levels of data protection that prevents not only against the loss of data, but also complies with regulations that ensure the necessary privacy.

Doing all this — and much more! — is a real challenge for any IT organization and there is no one-size-fits-all solution. And this is where IBM Storage comes into play with a broad portfolio of Spectrum Storage software-defined-storage solutions, as well as all flash array and tape solutions from which an IT organization can select the right mix and combination of products (as well as services) to meet their particular needs. All of IBM’s storage and storage software solutions embrace this move to the multicloud enterprise.

IBM has already done a lot in the multicloud arena. See https://mesabigroup.com/ibm-flashsystem-9100-the-importance-of-nvme-based-stora will ge-in-a-data-driven-multi-cloud-world/ for an illustration. We will further illustrate with three examples of adding features, function, and a new product model that have come out of the recent IBM Storage announcement. These involve three areas which feature prominently in the multicloud world — namely, modern data protection with IBM Spectrum Protect and IBM Spectrum Protect Plus, mainframe storage with the introduction of the DS8882F model, and cloud object storage with enhancements to IBM Cloud Object Storage to better manage large quantities of data.

What IBM Is Doing for Multicloud Data Protection

Data protection is often a prominent use case for the multicloud. For example, storing a backup copy on a public cloud may be more cost-effective than storing it on-premises; however, storing it in a public cloud might negatively impact an RTO (recovery-time objective). IBM Spectrum Protect 8.1.6 resolves that dilemma by creating a tiering option for backup data based upon state. Data in an active state (which means the most recent backup copy) remains on-premises to help meet RTO needs while inactive data (which means previous backup copies) are stored in a cloud to reduce costs.

IBM had previously announced “solution blueprints” to help IT organizations to more easily deploy to the multicloud for a particular purpose, such as modern data protection or reuse of secondary datasets. Now blueprints are available with IBM Spectrum Protect to make it easier to deploy to popular cloud environments, namely IBM Cloud, Amazon AWS, and Microsoft Azure.

While IBM Spectrum Protect focuses on traditional “real” environments, IBM Spectrum Protect Plus focuses on data recovery and reuse in virtual machine environments (such as those managed by VMware vSphere). IBM Spectrum Protect Plus has added encryption of vSnap data repositories, as well as support for vSphere 6.7 and DB2.

Mainframe Storage Plays a Key Role in the Multicloud World

IBM has announced the introduction of the DS8882F, the latest all-flash storage system in the DS8880F mainframe enterprise-class storage family. The DS8882F fits into the same 19-inch industry standard rack with the IBM mainframe, while delivering up to 50% savings in power consumption and in physical space as compared to deploying an array as a separate item on the data center floor, while delivering from 6.4 to 368.64 TB of raw capacity. It is the first enterprise class storage system that can be integrated into IBM Z model ZR1 or IBM LinuxONE Rockhopper II. It also provides a straightforward and cost-efficient means for the upgrade of legacy mainframe storage systems, such as the non-flash DS6000, DS8000, and DS8870 systems.

Along with its fellow members of the DS8880F family, the DS8882F plays in the multicloud world through Transparent Cloud Tiering (TCT). See https://mesabigroup.com/ibm-introduces-transparent-cloud-tiering-for-ds8880-storage-systems/ for an introduction to TCT for DS8880 systems. The D8880 family in conjunction with TCT now provides a Safeguarded Copy capability that protects sensitive data copies, as well as enhanced data encryption before a DS8880 family member sends data to a cloud or IBM Cloud Object Storage.

IBM Cloud Object Storage Support of the Multicloud

IBM Cloud Object Storage (COS) software offers IT organizations a wide range of cost-effective options on how to store vast quantities of data that range from more active data, such as required for ongoing analytics, to less-frequently-used colder data, such as for archiving or backup, on a broad set of multicloud platforms, including on-premises and the public cloud.

Although COS already supports over 30 hardware server configurations from Cisco, Dell, HPE, and Lenovo among others in addition to IBM, the hardware verification process for a new server could take months. But this verification process has been reduced to weeks. With the next verification process a Quanta system that can start as an entry level solution in 1U at each of three sites and grow online while never going down (due to the very high availability of a COS system) and even upgrade to exabyte scale (technically, as very few if any are likely to reach those exalted levels).

COS plays an important role in data protection (such as backup) and lifecycle management (such as archiving) using more than 77 certified applications. IBM Cloud Object Storage includes three other IBM solutions, namely, DS8880 arrays using TCT, IBM Spectrum Scale NAS storage, which also uses TCT, and IBM Spectrum Protect enterprise backup software.

Mesabi musings

Multicloud represents an powerful and flexible evolutionary view of cloud strategy that says that enterprises need to distribute their applications and data across multiple clouds. Changing an information infrastructure dramatically always presents a number of challenges to IT organizations. IBM brings to the table a broad range of software-defined-storage software and storage systems solutions to help IT organizations address those challenges.

IBM plays an important role in multicloud and that was demonstrated in 3 different areas in this paper. The first was modern data protection, which is critical for multicloud deployments. The second is mainframe storage, where IBM is one of the leading storage providers, and which must not be neglected in a multicloud solution for those organizations using mainframes. The third is IBM Cloud Object Storage, which provides a cost-effective and reliable means of storing data in the cloud. And these are only three of a number of solutions that IBM Storage is delivering to make the move to the multicloud real and viable.

IBM FlashSystem 9100: The Importance of NVMe-based Storage in a Data-Driven Multi-Cloud World

IBM’s newly announced FlashSystem 9100 is its first NVMe (nonvolatile memory express) at the storage drive-level storage system. The FlashSystem 9100 is IBM’s enterprise-class entrée in the virtual storage infrastructure managed by Spectrum Virtualize.

But the announcement is about more than just an array solution. The true value is in how the FlashSystem 9100 makes a major contribution to the multi-cloud worlds where IT organizations increasingly play.  The FlashSystem 9100 includes an extensive set of IBM’s award-winning Spectrum Storage software and leverages that included software to create multi-cloud solution blueprints for IBM clients and channel partners.

The NVMe storage revolution

NVMe is a storage technology that accentuates, accelerates and revolutionizes the move to all-flash storage systems, as it supports solid state devices (SSDs) and not hard disk drives (HDDs). As a review of NVMe basics and IBM’s commitment to NVMe, please see http://mesabigroup.com/ibms-strong-commitment-to-the-nvme-storage-revolution/. However, to summarize, each new generation of a high technology system typically brings with it price performance benefit increases in speeds and feeds. NVMe is no exception.

The IBM FlashSystem 9100 speeds and feeds

From simply a speeds and feeds perspective, the FlashSystem 9100 offers 6X more data in the same space, 3X more performance, and 60% less energy consumption than traditional all-flash arrays. The Spectrum Virtualize-managed FlashSystem 9100 uses IBM’s FlashCore architecture at the storage module level.

No Future Worry Capacity Planning with FlashSystem 9100 Capacity Increases

However, there is a major new twist; the FlashCore storage modules in the 9100 have been redesigned to use the industry standard 2.5-inch form factor instead of IBM’s proprietary 10-inch form factor. The 9100 also uses 3D TLC (3-dimensional triple level cell) NAND flash with 64 layers instead of the 32 layers of the previous version. Finally, each FlashCore module offers built-in, performance-neutral hardware compression and data encryption.

These upgrades offer significant practical value. IBM expressly guarantees at least a 2 to 1 data reduction ratio standard without requiring the customer to submit to any testing and will flexibly guarantee up to a 5 to 1 data reduction ratio if the customer agrees to allow testing to show that the better compression ratio will actually apply to the customer workloads. Data reduction techniques include not only compression, but also deduplication and thin provisioning.

As a result, a single 2U 9100 system can hold up to 2 PB (petabytes) of data, and a fully populated cluster in a standard 42U data center rack can hold up to 32 PB. That is a mammoth amount of data to store in a small space. Most customers will not have that much data even in the foreseeable future, but the point is that with the FlashSystem 9100, you never have to worry about running out of storage capacity again!

NVMe-based Acceleration Turbocharges FlashSystem 9100 Performance

All the benefits of NVMe at the device level translates into a 3X performance increase over traditional all flash products. The latency for a single 2U array or a 4-way 8U cluster is the same at 100 microseconds, but the IOPS quadruples from 2.5 M/sec to 10 M/sec, and the bandwidth quadruples from 34 GB/sec to 136 GB/sec. The 9100 is truly a turbocharged system.

The IBM FlashSystem 9100 Comes in Two Flavors

The IBM FlashSystem 9100 comes in two models — the FS9110 and the FS9150. The former uses dual 8-core processors per controller and the FS9150 uses dual 14-core processors per controller. Otherwise the architecture is the same with up to 24 bays full of dual-ported 2.5” NVMe flash-based storage modules in 2U. There is also a minimum of two controller canisters that act in an active-active mode with failover/failback capabilities. An IT organization has to decide which model based upon how heavily the controllers would be used for their specific workloads.

IBM FlashSystem 9100 data-driven, multi-cloud solutions

The FlashSystem 9100 is about much more than speeds and feeds, such as being NVMe-accelerated. IBM is also targeting the rapidly emerging multi-cloud world where businesses are deploying private, hybrid, and public clouds in various and diverse combinations.

IBM offers customers a choice of three IBM validated “blueprints” that they can utilize to aid them in delivering a particular multi-cloud solution.

  1. The Data Re-use, Protection, & Efficiency Solution focuses not only on how to backup data in virtual or physical environments such as using IBM Spectrum Protect Plus, but also how to re-use backup and other copies for DevOps, analytics, reporting and disaster recovery (DR), while also adding in the use of IBM Spectrum Copy Data Management.
  2. The Business Continuity and Data Re-use Solution focuses on how to use storage in the public IBM Cloud as a DR target with easy migration among on-premises, private cloud, and public cloud. IBM Spectrum Virtualize for Public Cloud is used in addition to IBM Spectrum Virtualize and IBM Spectrum Copy Data Management.
  3. The Private Cloud Flexibility and Data Solution focuses on delivering on-premises or private cloud storage with cloud efficiency and flexibility for Docker and Kubernetes environments for new generation applications. IBM Cloud Private and IBM Spectrum Access Blueprint are used in the deployment process.

IBM software-defined storage targets the multi-cloud world

Software is the integrating glue that ties the NVMe-accelerated IBM FlashSystem 9100 to the multi-cloud, as enterprise-class storage systems are not only about hardware. In addition to a wide range of data services, such as snapshots and data replication, IBM includes with each FlashSystem 9100 access to the IBM AI-based Storage Insights as well as integrating four key members of its storage software and modern data protection family of Spectrum Storage solutions: namely, Spectrum Copy Data Management, Spectrum Protect Plus, Spectrum Virtualize for Public Cloud and Spectrum Connect.

IBM Storage Insights is a powerful tool for managing storage that in addition to helping with event and problem resolution management also provides infrastructure planning capabilities for forecasting capacity growth, planning purchases and optimizing data placement.

As for the four Spectrum Storage products, Spectrum Copy Data Management provides ongoing visibility into where data lives, how that data is used and who has access to it through data lifecycle management automation that delivers self-service data access along with necessary orchestration and visibility features. Spectrum Protect Plus focuses on easy-to-use backup and recovery in virtual environments. Spectrum Virtualize for Public Cloud connects on-premises and cloud storage (private or public) in order to deliver a hybrid cloud storage data replication and disaster recovery solution. Spectrum Connect enables the provisioning, monitoring, automating and orchestrating of IBM block storage in containerized (Dockers and Kubernetes), VMware and Microsoft PowerShell environments.  

Now what do all of these software products have in common? The answer is that they are integrated with the FlashSystem 9100 storage architecture that includes support for key capabilities, such as data portability between private and public clouds, native DevOps capable, containerization support, and self-service-enabled, that go beyond traditional block-based applications.

IBM states that modern IT organizations face three major challenges in the multi-cloud world that these software products address in conjunction with the FlashSystem 9100 addresses. The first is the need to modernize traditional applications in private clouds, which bring the agility, flexibility, and cost effectiveness of a public cloud, while at the same time being able to extend seamlessly to and leverage public clouds as appropriate. The second is to be able to adopt successfully new data-driven applications, such as big data and a host of analytically-oriented applications. The third is the ability to modernize applications, such as containerization in private clouds that use agile development approaches with full portability that leverages the public cloud infrastructure. The multi-cloud world is here to stay and the FlashSystem 9100 has been designed to play effectively in that world.

The IBM FlashSystem passes the litmus tests of reliability and pricing with flying colors

All of the above attests to the power of the FlashSystem 9100, but IT organizations also want to know about issues, such as reliability and pricing. On the reliability front, IBM guarantees 100% data availability for users of HyperSwap, which is a Spectrum Virtualize capability that is used in a dual-site, active-active environment. In addition, IBM offers a seven-year life on the FlashCore media itself while on warranty or extended maintenance, which should end any concern over read/write endurance.

Pricing deals with many factors, such as total cost of ownership. However, IBM believes in current customer retention and providing enticements to attract new customers for whom the multi-cloud world presents challenges that the FlashSystem 9100 can solve. Therefore, the price of the current V9000 and a 9100 are roughly equivalent. For example, if a V9000’s warranty period has expired and an IT organization is willing to buy three years of maintenance support on the system, then it could acquire a new 9100 with its warranty for approximately the same price. Now, IT would have to migrate its data to the new array, but since Spectrum Virtualize is used on both systems, that data migration could be made non-disruptively. That is what is called a good deal.

Mesabi musings

What’s not to like about IBM’s new FlashSystem 9100? NVMe-accelerated performance, solid data reduction, starting with compression, multi-cloud functionality to deal with the world that IT organizations now must face more and more each day, and multiple PB capabilities to name just a few — and all this in only a 2U box!

As a side note, NVMe at the storage device level in all-flash systems drives a final nail in the coffin for the use of hard disks for Tier 1 production storage. On the positive side, enterprise-class NVMe storage is the way to go in the rapidly growing multi-cloud world and the IBM FlashSystem 9100 is a clear illustration of why that is the case.

IBM Storage Insights: Here’s To Your Storage’s Health — And More

Storage systems are inherently complex and IT users need to manage their storage environment’s performance, capacity utilization, and health constantly. Vendors have long helped with Call Home capabilities where a storage system sends storage usage data to a vendor. Now IBM has turbocharged Call Home with Storage Insights where more data is collected, where users are able to better self-service their needs through using a feature-rich dashboard, and where IBM can provide deeper and broader technical support when the user needs that extra level of storage management support. Let’s look more deeply into IBM Storage Insights.

IBM Storage Insights Delivers a Turbocharged Call Home Capability

Call Home has long been a standard and well-accepted feature for many block-based storage systems whereby metadata (such as on performance and capacity utilization) is transmitted from a customer datacenter to a vendor site for storage monitoring purposes. The data can then be used for diagnostic, analysis, and planning purposes that can include proactive alerts to avert a potential problem (such as an early detection of a bad batch of disks that are starting to degrade below acceptable levels) or to more rapidly accelerate the resolution of a problem that has unexpectedly occurred.

Although Call Home capabilities vary among vendors, traditional systems can be limited in a number of ways:

  • Reactive alerts only to error conditions such as hardware failures as limited metadata prevents broader usage value; for example, among many other concerns, this means that proactive support for configuration optimization may not be available
  • Users do not have an interface with the system at the vendor site that allows them to self-service, self-manage the process as much as possible; that means a greater (and unnecessary) level of reliance on the vendor for support; while necessary support is valuable, you do not want to in-effect delegate decision-making to someone who is not as familiar with your storage systems as you are
  • May focus on individual storage systems rather than on all the storage systems so there is no unified pane of glass for an IT user to view all critical events easily (usually at a single glance); this makes a storage administrator’s life more difficult

The overview of IBM Storage Insights below reveals how IBM turbocharges Storage Insights to overcome those limitations and to provide even more features and functionality.

Overview of IBM Storage Insights

IBM Storage Insights is software that runs on the IBM Cloud. A lightweight data collector installed in the user data center streams performance, capacity, asset, and configuration metadata to the IBM Cloud. This data is metadata because it is data about data rather than the actual applications data, which the data collector cannot touch. Metadata flows in only one direction, from the user data center to the IBM Cloud via HTTPS.

Among the types of metadata provided are:

  • Performance metrics, including I/O rates, read and write data rates, and response times
  • Capacity information, such as used space, unassigned space, and the compression ratio
  • Storage system resource inventory and configuration information, such as volumes, pools, disks, and ports

The metadata is made available to the Storage Operations Center at the IBM Cloud. The metadata is analyzed and presented in the form of a dashboard where those authorized can see the performance (such as I/O rate, data rate, and response time), health, capacity, and call home events for all IBM block storage systems from a single, unified pane of glass. Each storage system is displayed on the dashboard in a small rectangular slab called a tile.  Numerous tiles (say 28 depending upon screen size) can be displayed on a single screen before scrolling would be necessary; however, little or no scrolling is likely to be necessary as system tiles for devices with a critical event automatically reshuffle to the top of the screen and the user can also press a filter button to see only those events that require immediate attention. Users can also create custom dashboards showing a subset of storage systems. The user, say a storage administrator, can click on the tile with a critical event and drilldown to a pull-out drawer that provides more details to help diagnose the issue. This self-service diagnostic data can often help the user resolve the problem without having to request IBM support.

IBM follows a number of security practices in the IBM Cloud, including physical, organizational, access, and security controls. IBM Cloud is ISO/IEC 27001 Information Management Security certified. Access is restricted to authorized user company personnel as well as the IBM Cloud team that is responsible for the daily operation and maintenance of IBM Cloud instances and the IBM Support team that is responsible for investigating and closing service tickets.

IBM Storage Insights overcomes the limitations of traditional Call Home systems:

  • Although some problems cannot be prevented and a reactive approach is necessary, Storage Insights provides a trouble ticket process that can speed the resolution of a problem. However, Storage Insights enables proactive insights as well. For example, a health check may show only a single connection to a server from a storage system. Although everything is running fine now, this exposes a potential future problem that could be prevented by taking action now to add an extra connection.
  • The dashboard empowers storage administrators to control their own destiny better through its self-service capabilities,
  • The dashboard also enables a storage administrator to manage all IBM block storage as a whole that is much easier that trying to manage on a piecemeal basis.

In addition, IBM offers capabilities, features, and functions far beyond traditional Call Home approaches:

  • In addition to event and problem resolution management tool that tends to be the limit of traditional Call Home systems, IT can use infrastructure planning tools to forecast capacity growth, plan purchases, and optimize data placement; this means that IT can save money through better utilization of existing storage as well as making sure that they do not overinvest in making future storage system purchases.
  • IBM Support has a wealth of data, including historical data, and can use data scientists who fish the data in a secure data lake to do analyses such as using best practice thresholds to help identify anomalies, such as why an application’s performance suffers at unexpected times.
  • Being cloud-based means that IBM can add features and functions on the fly: everyone has access to the most current version.
  • The data is protected in accordance with GDPR.

IBM Storage Insights is available for free; IBM does this as it strengthens their relationship with their customers. A fee-based Pro version is available, but IT should become familiar with the standard version first before deciding if the Pro version (such as capturing longer periods of historical data for planning purposes) is necessary.

Mesabi musings

Managing a large, complex storage infrastructure has always been a difficult task. Not only do applications have to deliver the needed performance (such as response time), but they have to do so without running out of space (capacity) and without unexpected downtime (health). Moreover, IT wants to do this while using their storage resources most efficiently from an infrastructure investment management perspective (as there is no sense in paying more than necessary for desired levels of performance).

For IBM block storage users, the process has just gotten a lot easier with the introduction of IBM Storage Insights. IBM continues to put the IBM Cloud to good use as Storage Insights enables an effective self-service capability for the IT user through a unified view of all of its IBM storage through a single pane of glass along with a rich and robust set of IBM Support (such as the use of data scientists) services as necessary.

And since the price is free, no one who uses Call Home now should hesitate to adopt IBM Storage Insights. In addition to all flash storage systems and software-defined storage, the type of features and functions that IBM Storage Insights provides is likely to see a similarly high adoption rate because of the value that it can deliver in managing the health, capacity, and performance of a storage environment.