Storage protocols comparison – Fibre Channel, FCoE, Infiniband, iSCSI?

100Gbps
40Gbps
FCoE
Infiniband
iSCSI
SAN
EDGE Technologies
10 min read
Protocol comparison

Data is Growing:

The amount of digital data in the universe is growing at an exponential rate, doubling every two years, and changing how we live in the world. If we look at data, we can divide them into: Structured Data (highly organized and made up mostly of tables with rows and columns that define their meaning, like Excel spreadsheets and relational databases) and Unstructured Data (Books, letters, audio, video files, power presentations, images…)

Structured-Data-100

The volume of unstructured data has exploded in the past decade:

Data-Explosion-100

The growth of structured versus unstructured data over the past decade shows that unstructured data accounts for more than 90% of all data. “Between the dawn of civilization and 2003, we only created five exabytes; now we’re creating that amount every two days. By 2020, that figure is predicted to sit at 53 zettabytes (53 trillion gigabytes) — an increase of 50 times.” — Hal Varian, Chief Economist at Google. There is a good example of it even in traditional industries like Energy, as industry is changing. Below is a quote from General Electric's vice-president of the software center: “The amount of data generated by sensor networks on heavy equipment is astounding. A day’s worth of real-time feeds on Twitter amounts to 80GB. One sensor on a blade of a turbine generates 520GB per day, and you have 20 of them.” These sensors produce real-time big data that enables GE to manage the efficiency of each blade of every turbine in a wind farm.

Data needs to be stored:

As we have these large data amounts it becomes very important which technology we choose for storing the data. There are about 3 important parameters that will have a large impact on how fast we will access our stored data:
1. Bandwidth – a measure of the data transfer rate, MB/s;
2. Latency – time taken to complete an I/O request, also known as a response time. Usually measured in milliseconds (1/1000 of a second). Going forward in microseconds (1/1000 of a millisecond);
3. IOPS – Input/output operations per second (IOPS, pronounced eye-ops) is an input/output performance measurement used to characterize computer storage devices like hard disk drives (HDD), solid state drives (SSD), and storage area networks (SAN);

Two other parameters every data center cares about are: power consumption and reliability.

Storage protocol comparison – Fibre Channel, FCoE, Infiniband, iSCSI:

There are several types of storage protocols to choose from, and based on the choice will largely depend on our networking parameters, what type of network infrastructure we are going to have, even what brand switches and routers we are more likely to see in our data center, and what cabling type we will have in it.

Below would be the summary about possible choices and a bit deeper look at each one of them:

AttributeInfinibandFibre ChannelFCoEiSCSI
Bandwidth (Gbps)2.5/5/10/14/25/508/16/32/128 10/25/40/100 10/25/40/100
Adapter Latency25us50us200usWide range
Switch latency per hop100-200 ns700 ns200 ns200 ns
AdapterHCA – host channel adapterHBA – host bus adapterCNA – converged network adapterNIC – network interface card
Switch BrandsMellanox, IntelCisco, BrocadeCisco, BrocadeHPE, Cisco, Brocade
Interface Card BrandsMellanox, Intel Qlogic, EmulexQlogic, EmulexIntel, Qlogic
Deployment amountXXXXXXX
ReliabilityXXXXXXXXX
Ease of management XXXXXXXX
Future upgrade pathXXXXXXXXXX

*Please note that, as we didn’t have a chance to get the above adapter latency figures and switch latency per hop figures from a neutral source, they are a bit relative, but can give an approximate idea about each protocol's performance in this area.

x – Low
xx – Medium
xxx – High

InfiniBand (IB)

InfiniBand (IB) is one of the latest computer-networking communications standards used in high-performance computing that features very high throughput and very low latency, and it is most commonly used to interconnect supercomputers. Mellanox and Intel are two main manufacturers of InfiniBand host bus adapters and network switches. Intel jumped into this game by acquiring QLogic's InfiniBand division and Cray’s HPC interconnect business.
Speeds vary from 2.5Gbit/s (SDR) up to 50Gbit/s (HDR) and can also be combined in up to 12 links, where we can get a 600 Gbit/s connection. For more info, please see the table below:

SDRDDRQDRFDR10 FDREDRHDR NDRXDR
Adapter latency (microseconds)52.51.30.70.70.5
Signaling rate (Gbit/s)2.551010.3125 14.062525.7812550100250
Speeds for 4x links (Gbit/s) 816324054.54100200
Speeds for 8x links (Gbit/s) 16326480109.08200400
Speeds for 12x links (Gbit/s)244896120163.64300600
Year2001200520072011201120142017after 2020future

For Infiniband interconnection, we can use DAC cables, AOCs, and optical transceivers, and we can use DAC and AOC breakout cables.

InfinibandSpeedForm FactorDACAOCsOptical transceiversDAC breakoutAOC breakout
HDR200Gb/sQSFP56YYYYY
EDR100Gb/sQSFP28YYY
FDR56Gb/sQSFP+YYY
FDR10 QDR40Gb/sQSFP+ YYY
QDR40Gb/sQSFP+YYY
Distance0.5m to 7m3m to 300m30m to 10km1-2.5m3m to 30m

Fibre Channel (FC)

Fibre Channel (FC) became the leading choice for SAN networking during the mid-1990s. Traditional Fibre Channel networks contain special-purpose hardware called Fibre Channel switches that connect the storage to the SAN, plus Fibre Channel HBAs (host bus adapters) that connect these switches to server computers. Fibre Channel is a mature, low-latency, high-bandwidth, high-throughput protocol. As a storage protocol, FC is simple to configure and administer, and has had wide adoption over the past 18 years.
Fiber channel speeds have evolved, and nowadays 16GFC, 32GFC, and 128GFC speeds are supported in modern FC switches and HBAs, and 256GFC and 512GFC speeds are in the roadmap.

NameLine-Rate GigabyteAvailable Since
1GFC1.06251997
2GFC2.1252001
4GFC4.252004
8GFC8.52005
10GFC10.518752008
16GFC14.0252011
32GFC28.052016
128GFC28.05×4=112.22016
256GFC2021
512GFC2024

If we check, then mainstream brands for Fibre Channel switch technology are Cisco, Brocade, and mainstream brands for HBAs are QLogic and Emulex. There are also Dell and Lenovo FC switches, but these are Brocade OEM equipment.

For FC switch and HBAs interconnection, we use optical modules. Below would be very useful material for all Brocade FC switch users about the Brocade Fibre Channel Transceiver Support Matrix.

We can see that Brocade supports FC modules with SWL (Short Wavelength up to 300m), LWL (Long Wavelength, up to 10km), and ELWL (Extended Long Wavelength up to 25km). EDGE Optical Solutions can offer not only standard modules but also hybrid solutions, which will support longer distances (for example, up to 100km for 8G FC), and as we have 8G/16G FC CWDM and DWDM types, which can be very useful for data center interconnection. Another good interconnect resource if you have Cisco-based FC infrastructure: Cisco MDS 9000 Family Pluggable Transceivers Data SheetHere we can see the relationship between Cisco FC platforms and available Fibre Channel Transceivers for them.

Fibre Channel over Ethernet (FCoE)

An alternative form of FC called Fibre Channel over Ethernet (FCoE) was developed to lower the cost of FC solutions by eliminating the need to purchase HBA hardware. FCoE is FC over Lossless Ethernet.

Fibre Channel required three primary extensions to deliver the capabilities of Fibre Channel over Ethernet networks:

  • Encapsulation of native Fibre Channel frames into Ethernet Frames.
  • Extensions to the Ethernet protocol itself to enable an Ethernet fabric in which frames are not routinely lost during periods of congestion.
  • Mapping between Fibre Channel N_port IDs (aka FCIDs) and Ethernet MAC addresses.

The key advantage of FCoE is the ability to use Ethernet cabling for datacenter network and storage traffic, which removes the need for Fibre Channel fabric and infrastructure. It reduces the number of cables as well number of interface cards required. Instead of separate Ethernet network cards and Fibre Channel host bus adapters (HBAs), the two functions will be combined onto a single converged network adapter (CNA). The CNA deals with both protocols on a single card, while allowing storage and network domains to be controlled independently.

Now, many Fiber Channel switches support FCoE, for example, Cisco Nexus series and Cisco MDS series support it, as Brocade switches support it. Cisco was one of the vendors that was pushing FCoE hard, as Cisco was not as good as Brocade on FC, but it had a lot of expertise in Ethernet and IP. While FCoE promises improved economics, it does not improve availability or performance relative to FC. It does add administration complexity. Another consideration is the linkage between the low adoption rate and the lack of experienced FCoE administrators, service, and support personnel. Adoption of FCoE has not been successful, and as initially expected by FCoE vendors, FCoE has not replaced FC.

iSCSI

iSCSI was created as a lower-cost, lower-performance alternative to Fibre Channel and started growing in popularity during the mid-2000s. iSCSI works with Ethernet switches and physical connections instead of specialized hardware built specifically for storage workloads. It provides data rates of 10 Gbps and higher.
iSCSI appeals especially to smaller businesses that usually do not have staff trained in the administration of Fibre Channel technology. iSCSI was pioneered by IBM and Cisco in 1998. Unlike some SAN protocols, iSCSI requires no dedicated cabling; it can be run over existing IP infrastructure. As we use Ethernet more correctly, TCP, there is a large number of vendors, products, and components from which to choose. As a result, iSCSI is often seen as a low-cost alternative to Fibre Channel. However, the performance of an iSCSI SAN deployment can be severely degraded if not operated on a dedicated network or subnet (LAN or VLAN), due to competition for a fixed amount of bandwidth. For device interconnection in iSCSI, we can use 10G/40G/100G Ethernet optical modules, DACs, and AOCs.

Summary

Based on the above analysis, we can conclude that Fibre Channel is one of the best protocols for data storage network infrastructure, as it’s mature, low-latency, high-bandwidth, and it’s simple to configure and administer. As a downside, could are not many vendors from which to choose, which makes this technology quite expensive. This could be great technology for larger enterprises, as it is purpose-built to support storage, and that’s all it does.
iSCSI is also well good choice as a storage protocol as it is designed based on Ethernet, iSCSI is much less expensive than Fibre Channel. iSCSI uses standard Ethernet switches and cabling and operates at speeds of 1GB/s, 10GB/s, 40GB/s, and 100GB/s. Basically, as Ethernet continues to advance, iSCSI advances right along with it. Seems iSCSI was initially designed for smaller organizations that couldn’t afford a separate FC network and were willing to accept a bit lower performance for the sake of cost. In nowadays with Ethernet advancement, this could not be true anymore.
FCoE has not gone mainstream, also FCoE states their solutions are less complex and more cost-effective due to the reduction in cabling, along with the use of lower cost host adapters. FC can offer far better performance compared to FCoE.
Infiniband is a technology that can offer one of the best throughput and latency parameters, but the downside would be that it’s not so widely used, the administration part could be much harder than for other protocols, and from a cost perspective, it is an expensive technology.

These are our thoughts on the subject. As always, you’re welcome to comment, ask questions, or discuss: sales@edgeoptic.com

EDGE Technologies

Expert in telecommunications and data center technologies, sharing insights on the latest industry trends and innovations in optical networking solutions.