Data Storage Ebook
Data Storage Ebook
Managing
Growth
Data
In this issue:
Save Your IT Budget with Storage Tiering Expanding Your Storage Capacity Right-Sizing Your Backup Strategy
Issue 1, 2011
Contents
Data Storage Insights
Contributors: Drew Robb and Paul Rubens.
2 3
6 9
know when its time to expand their data storage capacity, including 10 signs that its time to upgrade your storage capacity. Finally, well examine backup strategies, which need to evolve as businesses grow so they can protect their data and recover from incidents to quickly resume normal IT operations. Well review the options that are available for businesses looking to right-size their backup strategy to ensure data protection.
Back to Contents
smart way for businesses to improve overall efficiency and save on data storage costs is by understanding the differences in the types of data they store. This can be achieved via storage tiering the process by which different types of data are stored in the way that makes the most sense for that data. Storage tiering, however, has long been thought of as the province of large enterprises due to the cost and complexity involved. This was perhaps true five years ago, but it is no longer the case. Tiering can now be accomplished economically even by small businesses. Storage tiering is also very important to midsize and growing businesses that are seeing the amount of data they store increase significantly and risk having their data storage infrastructure become overwhelmed. According to Mike Karp, principal analyst at Ptak, Noel & Associates, the purpose behind tiering is to arrive at an appropriate balance between storage cost and performance. Those embarking upon tiering, he said, have to come to terms with a common misunderstanding believing that all data is of equal value. Tiering involves the use of a number of different types of storage media (high performance/high cost at one extreme, and high-capacity/low-cost at the other end) and storage services (continuous backups for highvalue data, and weekly backups for lower value data, 3
for example), said Karp. This gives an IT manager the ability to assign proper performance and protection levels to data, which can be a great money saver. Savings are realized by assigning storage hardware and software services according to the value of the data. If all data is regarded as having a monotone level of importance, it might all be running on a top-of-the-line system and be afforded a formidable amount of data protection. Particularly in rapidly expanding companies, this soon becomes an expensive proposition. When data is analyzed, though, it becomes obvious that not all files are created equal. Almost every analysis reveals that perhaps 10 percent of the total is frequently accessed. This is almost always data less than 30 days old. After that, access rates diminish and beyond a certain points, files are rarely looked at, if at all. Karp points out, though, that there is more to it than time. Sales reports and accounts payable updates are examples of high-end data that might hold high value for a quarter or an entire year. Based on such analyses, storage tiering puts high-value material on high-performance hardware and assigns it the best backup and recovery services. Lesser value data, on the other hand, deserves proportionately less investment, and belongs on less expensive devices/services. When it comes to high-value data, back them up
Back to Contents
frequently, and make sure they are on the most reliable, high-performance storage systems in the shop, said Karp. Lower-priority data might include the corporate phonebook or the accounting departments football pool and should be assigned to cheaper storage that receives fewer services.
inactive data. Organizations can utilize tape, older SATA boxes or even the cloud. Note, too, that the above arrangements can vary from organization to organization.
Elements of Tiering
Traditional storage tiering consisted of speedy, but expensive Fibre Channel (FC) disks for Tier 1, cheaper disk such as SATA for Tier 2 and tape for Tier 3. But as technology has evolved and costs have plummeted, new tiering concepts have come on the market that can be implemented by small and midsized businesses. Solid State Drives (SSDs), for example, are frequently deployed as a Tier 0 to be used only by one or two mission-critical systems. There is Tier 0 using SSD where time is money or lives may be at stake, which is ultra fast but involves small amounts of space capacity, said Greg Schulz, an analyst with Server and StorageIO Group. However, it must be noted that SSD is many times more expensive than spinning disk. A clear value proposition has to exist if it is to be rolled out extensively. But inexpensive systems are coming on the market and they provide SSD capability at a reasonable cost. Schulz suggested that smaller companies that use databases may want to consider using SSDs but only for logs, tables and database metadata as this can provide a noticeable improvement in information retrieval rates even if the database file itself sits on SATA disk. Tier 1 can either be FC disk or less-expensive SAS drives that can operate at the same speed of 15,500 revolution per minute (RPM). Placing all information on this platform, though, isnt wise as this is the most costly form of disk. Tier 2 is generally composed of inexpensive SATA disk running at 7,200 RPM and available in sizes as large as 1TB or 2TB. Tier 3 is the lowest cost tier reserved for
Back to Contents
Schulz adds that time should be spent isolating inactive data and moving it off to a tape archive or inexpensive disk. This might consist of old databases or customer records from 10 years ago. Note, too, that tiering isnt only about hardware. Software is needed to track data patterns and move information automatically from one tier to another based on pre-set user parameters. Karp recommends, therefore, that organization investigate the software side before committing to a specific hardware platform.
Keep it Simple
Sabine Waterkamp, president of system integrator ACSLA Inc., advises IT organizations that are exploring storage tiering to keep it simple. Yes, three- or four-tiered systems may work for the largest of enterprises, but for those with less complex needs, two tiers might well be enough. For a typical SMB, I would suggest most often a 2-tier architecture using 15k RPM SAS for high transaction volumes and 7.2k RPM SATA for older data and backup storage, said Waterkamp. Unless you deal with large volumes of data, anything else than some sound placement of data on a 2-tier storage architecture may well be overkill.
Back to Contents
1. Your file servers are running out of physical space to add more hard drives.
Many companies rely on internal hard drives or direct attached storage (DAS) devices, but this type of storage cant be expanded indefinitely. Buying additional servers is not a sensible option because it is not a cost-efficient way to provide extra storage. It also adds administrative complexity and complicates backup procedures. A better solution could be to invest in networked storage can be shared by all of your existing servers.
2. Some of your storage resources are running out, while other resources are underused
If your company relies on internal storage or DAS devices that are only connected to one or a limited number of servers then you are probably using storage resources very inefficiently. Thats because you may have to invest in increased storage capacity on one server when you
Back to Contents
have unused capacity in another. More sophisticated options such as storage area networks (SANs) can help solve this problem.
6. Your local area network grinds to a halt during backups and when certain applications are busy
Applications that access large amounts of data using a legacy storage infrastructure can easily clog up your companys local area network, slowing data traffic to a crawl and reducing the productivity of all of your staff. This can be avoided if certain types of data are stored in a SAN and moved around on a fast dedicated storage network using a technology such as Fibre Channel.
3. Backing up is complicated and time consuming, and involves multiple backup systems
If you have expanded your existing storage capacity on an ad-hoc basis then its likely that you have many different types of storage devices, and groups of storage resource that are isolated from each other. This makes backing up complicated, time consuming and ultimately costly. If it becomes too complicated to be easily manageable then there is also an increased risk that your data is not all backed up properly, which can lead to catastrophic data loss. Centralizing your storage using a SAN array such as one from the Dell/EMC CX4 range combined with a Dell PowerVault tape-based backup system could make backup and recovery much more straightforward.
9. You have to buy more storage capacity than you need immediately
Adding capacity to your current systems may involve buying a large amount of additional storage that you will only use over a period of years. This is an inefficient
Back to Contents
and costly way of buying storage, because the utilization rate may initially be very low and a large proportion of this unused capacity may also require power, making your energy bills unnecessarily high. Newer storage systems let you take advantage of thin provisioning and overallocation. These two storage technologies enable you to prepare for storage capacity expansion while only buying storage media (such as hard drives) when they are required - instead of buying them up front and leaving then idle.
for short periods then you may have storage resources that are mostly unused. Newer storage systems enable you to benefit from storage cloud bursting making use of storage capacity in the cloud on a temporary basis and paid for per gigabyte per day. Insufficient data storage can bring a business to its knees, but organizations that need to add capacity need to be smart about the process. A number of relatively recent data storage technologies such as thin provisioning and data deduplication, when combined with strategies like storage tiering, will help businesses get the most for their storage dollars going forward so they can smartly and efficiently meet their demand.
10. You are paying for capacity you know you will rarely need
If your company occasionally has sudden, large and unpredictable storage capacity expansion requirements
Back to Contents
The key question to ask is whether the current storage backup and recovery strategy you have in place is still able to provide the level of data and business continuity protection that your business requires and can scale to meet your likely future needs, or whether you have effectively outgrown it.
Business Objectives
To determine if your current backup strategy can meet your business needs it may be useful to take some time to consider your recovery point objective (RPO) for each application or system. An RPO of one hour means that only data stored up to one hour before a disaster may be lost, but any data stored more than one hour before a disaster will have been safely backed up. Your RPO is therefore your possible data loss window. Deciding on a suitable RPO for your business begins by establishing a baseline the maximum data loss window that is acceptable. The other factor that needs to be considered is your recovery time objective (RTO) for each application or system. This is the maximum amount of time that it can take before your application or system is up and running again with restored data after a disaster. The RTO for each application or system will be determined by the cost to your business of that application or system being unavailable.
Back to Contents
RTO and RPO possible using tape is relatively long. The technology is proven and there is a wide range of backup software designed to work with automated tape backup systems. Pros Low per gigabyte storage costs Backup tapes can easily be moved offsite Tapes consumes no power when not in use Proven track record with many existing backup solutions Cons Slow read speeds and sequential (not random) data access means low RTO is not practical for large amounts of data Backup window can be long, making RPO relatively high Tape media can be hard to physically manage
Among the technology options that you may want to consider are:
10
Back to Contents
easy. The higher media costs associated with hard drives can be mitigated to some extent using data compression and deduplication, and data can easily be copied to tape media for offsite disaster recovery purposes or archiving. Pros Can be used with existing backup software and processes Faster and more reliable than tape systems Can take advantage of data compression and deduplication to reduce storage costs Cons Higher cost than B2T and D2D
Pros Zero RPO possible Captures every change to data as it happens so recovery is possible to any point in time Captured data can also be stored in secondary location for disaster recovery Very low RTO possible using high performance disks Systems can be recovered to any point in time Cons Costly Disk and network capacity intensive A thorough evaluation of the current backup strategy will help businesses understand if the strategy needs to be updated because of new requirements, more data or business growth. There are a number of options available for IT administrators that need to right-size their backup strategy, and each options carries with it some advantages and disadvantages.
11
Back to Contents