Get To Know The Storage I/O ChainStorage performance problems are often circular challenges. You fix one bottleneck and you expose another one. You can't really fix storage I/O, all you can do is get it to the point that people stop blaming storage for the performance problems in the data center. Getting there requires knowing the storage I/O chain.
Storage performance problems are often circular challenges. You fix one bottleneck and you expose another one. You can't really fix storage I/O, all you can do is get it to the point that people stop blaming storage for the performance problems in the data center. Getting there requires knowing the storage I/O chain.This is the sequence of components that start at the application and work their way down to the physical storage device. And this is a challenge automated tiering systems (ATS) face. These are solutions provided by vendors to attempt to solve storage I/O performance problems. They typically will move data based on the access frequency of that data. The more often the data is accessed the faster tier of storage the data is placed on, eventually landing on solid state disk (SSD).
The less accessed that data is the slower tier of storage that data is placed on, eventually landing on SATA based high capacity hard drives. There is little doubt that ATS will play an important role in the evolution of data center storage and the optimization of that resource. It is however just one component of the storage strategy, especially when it comes to performance.
Each component in the I/O chain needs to be measured and monitored to see if it can justify the investment that ATS and/or SSD are able to give it. Can the application generate enough simultaneous requests? Can the server process all those requests and get them on the NIC fast enough? Can that data travel across the connecting framework, through the switches maintain performance until it reaches the controllers in the storage system? Any break along this chain may obviate the value of ATS.
Ideally you want to upgrade just the right components to just the right level of performance to fix those issues. Determining what components should be upgraded and to what level requires tools to make those decisions. Interestingly most performance upgrades in data centers are more of a "cross your fingers and hope this fixes the problem" type of solution. Just throwing hardware at the problem leads to massive under-utilization and wasted resources.
Tools are needed that can monitor storage I/O performance from the application through the server (virtual or physical), through the HBA card, through the storage infrastructure and on to the storage system. This may even require physical tapping of the environment to get the exact performance benchmarks you need to make those decisions. While investments in these sorts of tools means an investment of precious IT budget dollars, when done as a first step it can avoid unnecessary upgrades and make sure that those upgrades you do implement will perform exactly as expected.
Track us on Twitter: http://twitter.com/storageswiss
Subscribe to our RSS feed.
George Crump is lead analyst of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. Find Storage Switzerland's disclosure statement here.