Hu Yoshida, Chief Technology Officer, Hitachi Data Systems, a wholly owned subsidiary of Hitachi, recently announced his top ten storage trends for 2011.
Hu Yoshida'ss Storage Predictions for 2011
- Storage Virtualization and Dynamic Provisioning acceptance will accelerate as it becomes the foundation for cloud and for dynamic, high availability data centers. Storage virtualization, the virtualization of external storage arrays, will provide the ability to non-disruptively migrate from one array to another and eliminate the costly down time required to refresh storage systems. Dynamic Provisioning enables storage to be provisioned in a matter of minutes, simplifing performance tuning with automatic wide striping, and enabling on demand capacity for an agile storage infrastructure.
- Closer integration of server and storage virtualization will be required to increase the adoption of data center virtualization. Server virtualization has matured beyond the cost reduction phase of consolidating print, file, test, and development servers and is currently poised to support tier 1 application servers. Moving forward, for support of tier 1 applications, server virtualization will need the integration of enterprise storage virtualization arrays that can offload some of the software I/O bottlenecks like SCSI reserves, and be able to scale to meet the high availability and QoS demands of enterprise tier 1 applications.
- Virtual tiering will be adopted for data life cycle management. Currently, virtual tiering has the ability to assign a volume to a pool of storage containing multiple performance, cost, tiers of storage and has the intelligence to move parts of that volume to different tiers based on access counts. The user does not need to classify a volume and assign it to a tier of storage, nor move the volume up and down the tiers based on activity. Virtual Tiering, or Dynamic Tiering, will do it automatically without the need to classify the volume and move the entire volume from tier to tier.
- The time is right for SSD acceptance for higher performance and lower cost in a virtual tiered configuration. Since 80% or more of a volume is usually not active, only a small amount of SSDs need to be in Tier 1 to serve the active parts of a volume while the majority of the volume can reside on lower cost SAS or SATA drives. A multi-tier storage pool that contains a small amount of SSD offset with a large amount of lower cost SAS and SATA drives could cost less than a single pool of SAS drives with the same total capacity and provide 4 to 5 times the IOPs.
- Serial Attached SCSI (SAS) will be adopted for increased availability and performance in enterprise storage systems. Unlike Fibre Channel (FC) loops, which are used to support FC drives on older storage systems, SAS is a point-to-point protocol. FC loops require each drive on the loop to arbitrate for access to the loop which causes contention. If a faster drive - like an SSD drive - is connected to the loop, it could drown out the loop so that the other drives could not get access. Since SAS drives are 6 Gbps and most FC loops are 4 Gbps, SAS has a performance advantage with its faster speed and point-to-point access. Since SAS is point-to-point, it is easier to identify a drive failure, as opposed to FC loops, which requires a query of each disk on the loop until the bad drive is found. SAS is also compatible with SATA. The only difference has to do with the ports - SAS is dual ported while SATA is single ported. In Hitachi storage arrays, SAS expanders are used as switches for the point-to-point connection. While IBM uses SAS drives in their DS 8800, they connect SAS drives through FC to their controllers. The drive vendors are quickly converting to SAS, for lower cost, performance and reliability.
- Small Form Factor Drives (SFF), will become prevalent for their power and cooling efficiencies. SFFs are 2.5 inch drives, which consume about 6 to 8 watts of power, as compared to Large Form Factor (LFF) 3.5 inch drives, which consume about 12 to 15 watts. This has a dramatic reduction in power and cooling, with an additional saving of floor space. Several vendors package 24 SFF disks in a drawer that is 2 U high and 33.5 inches wide. Hitachi changed the packaging on the AMS and the Virtual Storage Platform (VSP) so that the packaging is even denser. Instead of a drawer with all the drives mounted in the front, the AMS has a dense drawer with 48 drives that is 3 U high and 24 inches wide. The drawer pulls out for servicing with all 48 drives spinning. On the VSP, we have a disk module with 80 x 3.5 inch drives or 128 x 2.5 inch drives that is 13 U high and 24 inch wide. The disks are serviced from the front or from the back.
- Cloud will be accepted as a valid infrastructure model. Although some hype will still be associated with "cloud," there will be enough proof points to prove the concept. On-ramps to the cloud will facilitate the acceptance as well as management tools and orchestration layers that provide the end to end transparency to ensure service level objectives and chargeback.
- Convergence in the data center will begin to take off. The convergence of server, storage and network infrastructure will make it simpler and faster to deploy applications. The use of server, hypervisor, storage, and network virtualization will be key to providing an open platform to ensure investment protection and customer choice.
- Increased application transparency into a storage virtualization or cloud infrastructure will be required by applications. Without this transparency, application users will not be able to know if their service level objectives are being met, how to determine chargeback, how to plan their utilization, or the health of their infrastructure. Management software should provide a business unit or application dashboard in which an SLO is defined and persisted across configuration changes. The dashboard should show the status of the SLO, the actual allocation in terms of disk, RAID types, and storage ports, the health of the array groups and host links, and utilization of the allocated capacity over a selectable time frame.
- Remote managed services will be provided to offload the lower level monitoring, alerting, reporting, and management tasks that are limiting IT operations from moving to new technologies. For the past 10 years, the mandate for IT has been to do more with less, and operations staffs are overworked just to maintain more of the same. In order to transform the data center, the IT staff must find the time to train, plan and execute. A group of IT experts operating out of a Service Operations Center using remote management tools can leverage their skills across multiple installations at a very reasonable cost and drive higher and quicker return on asset investments.