San pioneer: start small, but smart


This article can also be found in the Premium Editorial Download "Storage magazine: Best storage products of the year 2002."

Download it now to read this article plus other related content.

I don't recommend the daisy-chain approach in general. Spend the money, buy more switches and design an architecture that will allow the availability, performance and flexibility necessary when SANs become larger and need to scale. For us that meant trunking ISLs when possible as well as building out core-edge topologies that scaled. This also let us take advantage of storage resources by merging fabrics when necessary, without having to schedule downtime. This approach also allowed us to build on our earlier investment in smaller switches - we introduce newer, faster, larger switches at the core and push the older, smaller ones to the edge or to development environments in some cases.

Requires Free Membership to View

Scaling backup
Backups are key to ensuring data integrity, and yet one of the most unrewarding jobs. Don't get boxed in with a tape technology, tape library or network that won't allow you to scale. Avoid practices that won't scale either, such as extensive use of homegrown scripts.

Tapes. Choose a tape technology that will allow you to protect your current investment in media. Although this may not be entirely possible, think in terms of flexibility and the business requirements. Media can be a huge cost for the company - when your company has over $1 million invested in DLT tapes, it's not easy to justify switching to a new tape technology that's not backwards compatible to your existing media, as we found out now that we want to graduate from DLT.

If you're using a specific media for your backup and recovery make sure that a plan is in place to migrate to the faster media as the drive speeds increase. Switching to faster media will also allow you to increase the size of your SAN without having to add more tape drives or tape libraries to accommodate the increased storage - which can be big expense to purchase and maintain.

Libraries. Select a tape library that will allow you take advantage of new tape drives and media. Your library should be flexible and will hold more than just one type of drive and media. It's more cost effective to buy a tape library that has the capacity to hold more drives than originally purchased than to order a tape library fully populated with drives. In the latter scenario you would be forced to upgrade the drives to a faster technology, which may require abandoning existing media at a huge loss for the company, or make another tape library purchase. Purchasing a larger library that's not fully utilized may cost a little more up front, but having the ability to add more drives, either same or faster, and mix the media within the library could potentially save thousands of dollars down the road.

You might have to add an additional library to the SAN, which may or may not be possible depending on your SAN design. Adding more tape libraries also increases the management complexity by requiring you to balance the backups over several tape libraries.

Design. How you design your backup and recovery depends on the methodology chosen.We chose client-free backups, which to us means we would use third mirror copies that could be split off and made visible to the backup server, which is connected to the external media device.

Even though there are several ways to design and configure a backup architecture, it is important that with whatever the decision, it will be inline with the three-year corporate strategy. By better aligning technical designs with the business strategy for the company several future costs can be avoided: The people costs for data migrations and management, the capital expenditures for new network, license and hardware purchases.

While it is possible to build out a core-edge topology using smaller switches, it's more cost-effective to use bladed directors with a large port count if high availability and increased flexibility are the goal. However, whether using smaller switches or large director switches to build your core-edge topology, the key is to pick an architecture that will scale for your environment.

While architecture is crucial, you can also avoid future costs by building out the right SAN infrastructure and how you connect servers to storage resources.

Start out small, and run fiber under the floor for a few hosts. But as you grow and have to move, add, or retire servers, switches, tape libraries and storage arrays, maintaining the fiber under the floor can be cost prohibitive. Troubleshooting a connection problem can be difficult with so many strands running on top of each other in a spaghetti fashion. Labels can become inaccurate or removed altogether from handling and moving the cables so many times.

If you don't have an infrastructure in place, you could end up with a lot of wasted fiber under the floor that could involve downtime to pull. Build the storage network like any other network and include patch panels with bundled fiber running between and distributed throughout the data center(s) in a design where the anticipated length from any server, storage array, switch and tape library can be calculated and preordered.

Understand the soft costs
One of the biggest soft costs when implementing and administrating a SAN are the people. This is probably because companies such as Intuit are telling vendors that such a tool is needed, a number of vendors are working on ways to easily administer the SAN from a centralized location and eliminate the manual bookkeeping tasks.

As our SAN grew from 20TB to 50TB to 200TB and over 900 switch ports - the old ways of managing were no longer practical. The spreadsheets we started with that detailed how the zones and disks were configured to the hosts wouldn't scale effectively. Again, this worked initially but keeping the documents updated always seemed to take a back seat to keeping the trains running.

Choose a SAN management tool early in your SAN deployment to cope with future growth. Good tools have widespread benefits in the area of interoperability, planning, scaling, and space reclamation (see "The right stuff").

If there's one thing I've learned from our whole experience, it's that basic technologies change rapidly. Absorbing them while running a real environment means you have to have good policies, procedures, design and management tools - so don't wait too long until you do that.

Now that we've arrived at a scalable architecture and infrastructure, we know what we need for management tools, and implementation is our biggest challenge. We're currently focusing on three main pain points. Implementing an enterprise resource management tool, better and more efficient ways for data archiving and remote data facility replication for achieving ensuring higher levels of data integrity and usability in the event of a disaster.

Change is still the order of the day. But I believe that with the current state of the art and the lessons that pioneers such as Intuit have learned, many companies can start at a reasonable level and grow into the many terabyte range while preserving most of their initial investment.

Web Bonus:
Online resources from SearchStorage.com: "The science of SANs," by M.C. Kinora.

This was first published in January 2003

There are Comments. Add yours.

TIP: Want to include a code block in your comment? Use <pre> or <code> tags around the desired text. Ex: <code>insert code</code>

REGISTER or login:

Forgot Password?
By submitting you agree to receive email from TechTarget and its partners. If you reside outside of the United States, you consent to having your personal data transferred to and processed in the United States. Privacy
Sort by: OldestNewest

Forgot Password?

No problem! Submit your e-mail address below. We'll send you an email containing your password.

Your password has been sent to: