Everyone seems to want to jump into purchasing a SAN, sometimes quite passionately. SANs are, admittedly, pretty cool. They are one of the more fun and exciting, large scale hardware items that most IT professionals get a chance to have in their own shop. Often the desire to have a SAN of ones own is a matter of “keeping up with the Jones” as using a SAN has become a bit of a status symbol – one of those last bastions of big business IT that you only see in a dedicated server closet and never in someone’s home (well, almost never.) SANs are pushed heavily, advertised and sold as amazing boxes with internal redundancy making them infallible, speed that defies logic and loaded with features that you never knew that you needed. When speaking to IT pros designing new systems, one of the most common design aspects that I hear is “well we don’t know much about our final design, but we know that we need a SAN.”
In the context of this article, I use SAN in its most common context, that is to mean a “block storage device” and not to refer to the entire storage network itself. A storage network can exist for NAS but not use a SAN block storage device at all. So for this article SAN refers exclusively to SAN as a device, not SAN as a network. SAN is a soft term used to mean multiple things at different times and can become quite confusing. A SAN configured without a network becomes DAS. DAS that is networked becomes SAN.
Let’s stop for a moment. SAN is your back end storage. The need for it would be, in all cases, determined by other aspects of your architecture. If you have not yet decided upon many other pieces, you simply cannot know that a SAN is going to be needed, or even useful, in the final design. Red flags. Red flags everywhere. Imagine a Roman chariot race with the horses pushes the chariots (if you know what I mean.)
It is clear that the drive to implement a SAN is so strong that often entire projects are devised with little purpose except, it would seem, to justify the purchase of the SAN. As with any project, the first question that one must ask is “What is the business need that we are attempting to fill?” And work from there, not “We want to buy a SAN, where can we use it?” SANs are complex, and with complexity comes fragility. Very often SANs carry high cost. But the scariest aspect of a SAN is the widespread lack of deep industry knowledge concerning them. SANs pose huge technical and business risk that must be overcome to justify their use. SANs are, without a doubt, very exciting and quite useful, but that is seldom good enough to warrant the desire for one.
We refer to SANs as “the storage of last resort.” What this means is, when picking types of storage, you hope that you can use any of the other alternatives such as local drives, DAS (Direct Attach Storage) or NAS (Network Attached Storage) rather than SAN. Most times, other options work wonderfully. But there are times when the business needs demand requirements that can only reasonably be met with a SAN. When those come up, we have no choice and must use a SAN. But generally it can be avoided in favor of simpler and normally less costly or risky options.
I find that most people looking to implement a SAN are doing so under a number of misconceptions.
The first is that SANs, by their very nature, are highly reliable. While there are certainly many SAN vendors and specific SAN products that are amazingly reliable, the same could be said about any IT product. High end servers in the price range of high end SANs are every bit as reliable as SANs. Since SANs are made from the same hardware components as normal servers, there is no magic to making them more reliable. Anything that can be used to make a SAN reliable is a trickle down of server RAS (Reliability, Availability and Serviceability) technologies. Just like SAN, NAS and DAS, as well as local disks, can be made incredibly reliable. SAN only refers to the device being used to serve block storage rather than perform some other task. A SAN is just a very simple server. SANs encompass the entire range of reliability with mainframe-like reliability at the top end to devices that are nothing more than external hard drives – the most unreliable network devices on your network – on the bottom end. So rather than SAN meaning reliability, it actually offers a few special cases of being the lowest reliability you can imagine. But, for all intents and purposes, all servers share roughly equal reliability concerns. SANs gain a reputation for reliability because often businesses put extreme budgets into their SANs that they do not put into their servers so that the comparison is a relatively high end SAN to a relatively budget server.
The second is that SAN means “big” and NAS means “small.” There is no such association. Both SANs and NASs can be of nearly any scale or quality. They both run the gamut and there isn’t the slightest suggestion from the technology chosen whether a device is large or not. Again, as above, SAN actually can technically come “smaller” than a NAS solution due to its possible simplicity but this is a specialty case and mostly only theoretical although there are SAN products on the market that are in this category, just very rare to find them in use.
The third is that SAN and NAS are dramatically different inside the chassis. This is certainly not the case as the majority of SAN and NAS devices today are what is called “unified storage” meaning a storage appliance that acts simultaneously as both SAN and NAS. This highlights that the key difference between the two is not in backend technology or hardware or size or reliability but the defining difference is the protocols used to transfer storage. SANs are block storage exposing raw block devices onto the network using protocols like fibre channel, iSCSI, SAS, ZSAN, ATA over Ethernet (AoE) or Fibre Channel over Ethernet (FCoE.) NAS, on the other hand, uses a network file system and exposes files onto the network using application layer protocols like NFS, SMB, AFP, HTTP and FTP which then ride over TCP/IP.
The fourth is that SANs are inherently a file sharing technology. This is NAS. SAN is simply taking your block storage (hard disk subsystem) and making it remotely available over a network. The nature of networks suggests that we can attach that storage to multiple devices at once and indeed, physically, we can. Just as we used to be able to physically attach multiple controllers to opposite ends of a SCSI ribbon cable with hard drives dangling in the middle. This will, under normal circumstances, destroy all of the data on the drives as the controllers, which know nothing about each other, overwrite data from each other causing near instant corruption. There are mechanisms available in special clustered filesystems and their drivers to allow for this, but this requires special knowledge and understanding that is far more technical than many people acquiring SANs are aware that they need for what they often believe is the very purpose of the SAN – a disaster so common that I probably speak to someone who has done just this almost weekly. That the SAN puts at risk the very use case that most people believe it is designed to handle and not only fails to deliver the nearly magic protection sought but is, to the contrary, the very cause of the loss of data exposes the level of risk that implemented misunderstood storage technology carrier with it.
The fifth is that SANs are fast. SANs can be fast; they can also be horrifically slow. There is no intrinsic speed boost from the use of SAN technology on its own. It is actually fairly difficult for SANs to overcome the inherent bottlenecks introduced by the network on which they sit. As some other storage options such as DAS use all the same technologies as SAN but lack the bottleneck and latency of the actual network an equivalent DAS will also be just a little faster than its SAN complement. SANs are generally a little faster than a hardware-identical NAS equivalent, but even this is not guaranteed. SAN and NAS behave differently and in different use cases either may be the better performing. SAN would rarely be chosen as a solution based on performance needs.
The sixth is that by being a SAN that the inherent problems associated with storage choices no longer apply. A good example is the use of RAID 5. This would be considered bad practice to do in a server, but when working with a SAN (which in theory is far more critical than a stand alone server) often careful storage subsystem planning is eschewed based on a belief that being a SAN that it has somehow fixed those issues or that they do not apply. It is true that some high end SANs do have some amount of risk mitigation features unlikely to be found elsewhere, but these are rare and exclusively relegated to very high end units where using fragile designs would already be uncommon. It is a dangerous, but very common practice, to take great care and consideration when planning storage for a physical server but when using a SAN that same planning and oversight is often skipped based on the assumption that the SAN handles all of that internally or that it is simply no longer needed.
Having shot down many misconceptions about SAN one may be wondering if SANs are ever appropriate. They are, of course, quite important and incredibly valuable when used correctly. The strongest points of SANs come from consolidation and special types of shared storage.
Consolidation was the historical driver bringing customers to SAN solutions. A SAN allows us to combine many filesystems into a single disk array allowing far more efficient use of storage resources. Because SAN is block level it is able to do this anytime that a traditional, local disk subsystem could be employed. In many servers, and even many desktops, storage space is wasted due to the necessities of growth, planning and disk capacity granularity. If we have twenty servers each with 300GB drive arrays but each only using 80GB of that capacity, we have large waste. With a SAN would could consolidate to just 1.6TB plus a small amount necessary for overhead and spend far less on physical disks than if each server was maintaining its own storage.
Once we begin consolidating storage we begin to look for advanced consolidation opportunities. Having consolidated many server filessytems onto a single SAN we have the chance, if our SAN implementation supports it, to deduplicate and compress that data which, in many cases such as server filesystems, can potentially result in significant utilization reduction. So out 1.6TB in our example above might actually end up being only 800GB or less. Suddenly our consolidation numbers are getting better and better.
To efficiently leverage consolidation it is necessary to have scale and this is where SANs really shine – when scale but in capacity and, more importantly, in the number of attaching nodes become very large. SANs are best suited to large scale storage consolidation. This is their sweet spot and what makes them nearly ubiquitous in large enterprises and very rare in small ones.
SANs are also very important for certain types of clustering and shared storage that requires single shared filesystem access. These is actually a pretty rare need outside of one special circumstance – databases. Most applications are happy to utilize any type of storage provided to them, but databases often require low level block access to be able to properly manipulate their data most effectively. Because of this they can rarely be used, or used effectively, on NAS or file servers. Providing high availability storage environments for database clusters has long been a key use case of SAN storage.
Outside of these two primary use cases, which justify the vast majority of SAN installations, SAN also provides for high levels of storage flexibility in making it potentially very simple to move, grow and modify storage in a large environment without needing to deal with physical moves or complicated procurement and provisioning. Again, like consolidation, this is an artifact of large scale.
In very large environments, the use of SAN can also be used to provide a point a demarcation between storage and system engineering teams allowing there to be a handoff at the network layer, generally of fibre channel or iSCSI. This clear separation of duties can be critical in allowing for teams to be highly segregated in companies that want highly discrete storage, network and systems teams. This allows the storage team to do nothing but focus on storage and the systems team to do nothing but focus on the systems without any need for knowledge of the other team’s implementations.
For a long time SANs also presented themselves as a convenient means to improve storage performance. This is not an intrinsic component of SAN but an outgrowth of their common use for consolidation. Similarly to virtualization when used as consolidation, shared SANs will have a nature advantage of having better utilization of available spindles, centralized caches and bigger hardware than the equivalent storage spread out among many individual servers. Like shared CPU resources, when the SAN is not receiving requests from multiple clients it has the ability to dedicate all of its capacity to servicing the requests of a single client providing an average performance experience potentially far higher than what an individual server would be able to affordably achieve on its own.
Using SAN for performance is rapidly fading from favor, however, because of the advent of SSD storage becoming very common. As SSDs with incredibly low latency and high IOPS performance drop in price to the point where they are being added to stand alone servers as local cache or potentially even being used as mainline storage the bottleneck of the SANs networking becomes a larger and larger factor making it increasingly difficult for the consolidation benefits of a SAN to offset the performance benefits of local SSDs. SSDs are potentially very disruptive for the shared storage market as they bring the performance advantage back towards local storage – just the latest in the ebb and flow of storage architecture design.
The most important aspect of SAN usage to remember is that SAN should not be a default starting point in storage planning. It is one of many technology choices and one that often does not fit the bill as intended or does so but at an unnecessarily high price point either in monetary or complexity terms. Start by defining business goals and needs. Select SAN when it solves those needs most effectively, but keep an open mind and consider the overall storage needs of the environment.
The latter part of this provides the real meat. The first part is a little condesecending, presumptuous, untrusting, and disrespectful of IT as a discipline. Quite a few typos that could be fixed with simple grammar check, by the way.
This was very enlightening 🙂
@Butch:
While I can see how you can take it that way, the first half does address the reality that many IT professionals (or the company’s non-technical “IT guy”) are often asked to perform projects they simply are not equipped to do. SAN and VMware are often sold hand in hand by vendors, and marketing reflects this. Even we are not immune to the latest fad or prestige of “running a SAN”.
I’m currently writing a whitepaper on SMB iSCSI SAN and this blog has been very helpful to explain when exactly a SAN is appropriate. Kudos to the author!
Sorry, I certainly disliked the tone of this post. I’m going to put this aside for a second to make some blunt comments.
1. I would not trade shared storage to go back to DAD (direct attached disk) ever.
I have been there and don’t want to go back. Shared storage has allowed me to reduce footprint and stop waste. Power, heat and space to start. Ever hear of server sprawl?
2. Maybe if your world doesn’t require TB and TB of storage you may not need a SAN… But In my opinion once you hit Medium biz you should be needing some sort of shared storage. If you are a small biz with you may still need shared storage.
3. Why hate on shared storage? I don’t get it, with all of the benefits of shared storage that make my job easier I can’t fathom why one would bash it. Oh sure, some solutions are too expensive but their are ways to fix that problem. I’ve done it so I know.
4. Yes, IT pros want the “Big boy” toys. I’ve seen folks had over hard earned money to Storage Vendors before and I likely will continue to see this. But this is not a reason to think you don’t need a SAN. If you don’t need a SAN you should know this already.
5. Virtualization likely requires a SAN. Now Microsoft has done something that is interesting and that’s the ability to Live-Migrate (vmotion for you ESXi only folks) in a shared nothing infrastructure. This is really the 1st stone I might be willing to stand on if your so Anti-SAN. But…. Your going to be buying a lot of hard disks because you need to have room to put your VM’s on each host you want to move to. Your also likely going to start looking at windows clusters for everything you care about… So your going to eat up Microsoft licenses fast.
6. Let me ADD to my 5th comment: your just shifting your spending $$ to other places. Your going to have more servers with more disk and RAM and to accomplish the same HA your going to waste $ on disk and RAM not is use if you expect to provide an environment that stays up. Your going to see a higher power bill and your going to have to spend more money on physical infrastructure. Your going to spend more on network just to plug everything in.
Look, if you have less than 50 servers you likely can get away without buying a SAN. But at some point you should realize you could do better with one(or many) and less “STUFF.” How many servers before you say you need it I’m not going to say. Why? Because their is more at play here, HA is about the total environment and that may be out of the scope of this author. DR and Business Continuity usually has Shared Storage as part of the equation for some specific reasons. One to mention is the WAN. Most companies hate to talk about WAN costs. Some try’s and use site to site VPN and large internet pipes. (Start at 100meg and work your way up) but when you have TB’s of stuff to protect that isn’t going to cut it. So you buy a shared disk solution to add on the storage replication solution you can afford.
Please understand there are a lot pieces in play here. The author should really point who the audience is here. I can tell you it’s not me nor the companies I work(Ed) for and it certainly doesn’t seem relevant to the companies I will likely work for in the future.
This post may fit for the SMALL biz but it really fits for the SMALL IT shop. That’s not where I want to be. I did that and I don’t want to go back.