HyperSCSI Examined 179
An anonymous reader writes "Eugenie Larson of byteandswitch.com has published a brief article that reviews the HyperSCSI protocol, which like iSCSI allows for an IP based san. The twist of HyperSCSI is that it's opensource, and runs over raw ethernet, avoiding the overhead of TCP/IP. The article has some comments from early adopters of HyperSCSI, as well as some comments from top vendors in the iSCSI industry."
Performance over TCP/IP? (Score:2, Informative)
If it's raw ethernet, then it's not "IP based" (Score:5, Insightful)
Which one?
Can't be both IP based and raw ethernet at the same time.
You don't expect us to RTFA do you?
Re:If it's raw ethernet, then it's not "IP based" (Score:1)
It does, however, call the protocol a "beer can with an engine" (or some such colorful metaphor for 'kludge').
I'm not a bit twiddler or an electrical engineer, but it looks to me like this is reinventing the wheel.
Re:If it's raw ethernet, then it's not "IP based" (Score:2, Informative)
Not quite reinventing, just reengineering. To keep the analogy, it's like reformulating the rubber to provide better "grip" in racing tires - great for flat, dry tracks, but not great for inclement weather. In this case, they are redesigning TCP to remove all of the stuff that is unneccesary for this particular purpose:
Re:If it's raw ethernet, then it's not "IP based" (Score:4, Informative)
You can have "without the overhead of TCP/IP" and "IP based". All IP gets you is an address format and ARP type standards, it's not a lot of overhead.
Re:If it's raw ethernet, then it's not "IP based" (Score:3, Insightful)
Re:If it's raw ethernet, then it's not "IP based" (Score:3, Informative)
IP is Layer 3. HyperSCSI is Layer 3.
Re:If it's raw ethernet, then it's not "IP based" (Score:4, Funny)
# # ###### ####
# # # #
# ##### ####
# # #
# # # #
# ###### ####
"HyperSCSI runs as a layer 3 protocol over Ethernet's layer 2."
Okay, so where's the IP layer? Wait, wait, don't tell me... it's on the bongos, right?
HyperSCSI runs on top of a raw datalink. IP doesn't enter into it.
Re:If it's raw ethernet, then it's not "IP based" (Score:2)
Re:If it's raw ethernet, then it's not "IP based" (Score:2)
It's pretty much tunneling Ethernet frames over TCP/IP.
Re:If it's raw ethernet, then it's not "IP based" (Score:2)
Its seems really strange for HS/IP to include the ethernet header and i get the impression from the PDF on their site that this isnt the case. Can you provide a reference?
Terminology again... (Score:2)
It does not at all mean that they are using TCP itself.
Re:If it's raw ethernet, then it's not "IP based" (Score:3, Informative)
Ethernet is both a Layer 1 topology and a Layer 2 Datalink protocol. That's why you can push ethernet frames over dissimilar topologies (Like 100baseFX and LANE over ATM).
OSI Layer 1 is Physical (Ethernet is here)
OSI Layer 2 is Datalink (Ethernet is also here)
OSI Layer 3 is Network (IP and HyperSCSI live here)
OSI Layer 4 is Protocol (TCP, UDP and the SCSI side of HyperSCSI live here)
Thanks for the lecture. (Score:3, Insightful)
If this uses it's own Layer 3 protocol (presumably, and for the sake of argument, called HyperSCSI), then it's NOT IP BASED... and the article summary indicated it was IP based, then contradicted itself.
Terminology. (Score:2)
So, if you say something is not based on tcp/ip,that indeed DOES mean it does not use IP. Furthermore, saying it uses "raw ethernet" indicates that this uses it's own layer 3 protocol, other than IP.
Re:If it's raw ethernet, then it's not "IP based" (Score:3, Interesting)
For people who wouldn't know this kind of stuff, TCP does much to ensure that every packet arrives as it was sent. This adds overheard, but it's hardly ever seen by any end user because it's pretty universal. UDP has no error checking, so it isn't fit for anything where any particular packet matters. On the plus side, overhead is severely reduced. I imagine UDP is used for streaming audio and video, but I don't know.
Re:If it's raw ethernet, then it's not "IP based" (Score:3, Insightful)
UDP (Score:2)
Re:If it's raw ethernet, then it's not "IP based" (Score:4, Informative)
For what they want UDP offers nothing that straight ethernet frames don't. And UDP has more overhead.
Re:If it's raw ethernet, then it's not "IP based" (Score:2)
RTFA? yes, perhaps you should have but instead you're moderated as insightful. (meta-mods?)
Sorting out the confusion. (Score:2)
Bridge Board (Score:4, Interesting)
Re:Bridge Board (Score:1)
According to the homepage for HyperSCSI, it can support IDE (as well as USB and Fibre Channel) devices:
Re:Bridge Board (Score:2)
Re:Bridge Board (Score:2)
But no cheaper than anything else could be if it gained popularity.
The big problem is the interrupts, and the processing power used. High speed transfers, and you need an incredibly fast computer to handle the storm of interrupts from your ethernet card. That's why (expensive) fibre channel is used, instead of (cheap) networking technologies.
Gigabit Ethernet (Score:2)
Re:Gigabit Ethernet (Score:2)
...and it won't be long before terabit ethernet is on the market to continue the revenue stream for our overlords.
Re:Bridge Board (Score:2, Insightful)
Reece,
favorite quotes (Score:1)
and
Mmmm
Re:favorite quotes (Score:2, Interesting)
I certainly appreciate his IDE efforts, but of course he is going to criticise the technology - his company is an iSCSI company!
What, do they think he is going to say, "Gee, and all this time, I've thinking that iSCSI is the right thing to work on. I'm going to abandon iSCSI right now, and start playing with this HyperSCSI thing."
It does implement reliability (Score:4, Informative)
The only technical negative side I can (at this time) is that because the implementation isn't over IP, you can't traverse a router. This usually isn't a problem but could cause some inflexibility in larger deployments.
Re:It does implement reliability (Score:1)
The proposed solution: (Score:3, Insightful)
That way you only need to buy one TOE card per WAN edge. Those can get expensive!
This isn't a problem with FC (Score:2, Insightful)
I don't know how far away you want to put your off-site backup, but Cisco have been selling a GBIC (Gigabit Interface Converter ? Too many FLAs for my head these days), which they've been calling 1000BaseZX, which will send an GigE signal around 90 Kilometers over single mode fibre.
Even Full Duplex Fast Ethernet over multi-mode fibre will go 2 Kilometers.
You can build some really big ethernet networks the
Re:It does implement reliability (Score:2)
Re:It does implement reliability (Score:2)
Re:It does implement reliability (Score:2)
This isn't a replacement for NFS or CIFS/SMB, but for FC and iSCSI.
No error checking or recovery?? (Score:2, Insightful)
And this is a technology breakthrough? I wouldn't want my data travelling down a wire with no error recovery no matter how small the error rate.
Reliability (Score:2)
Re:Reliability (Score:1)
Re:Reliability (Score:2)
My experience is that a properly installed and tested Ethernet network is very reliable.
Ha! (Score:1)
An therein lies your problem. You just can't get around the human error factor. People incorrectly install equipment and configuration mistakes abound in a corporate environment.
Re:No error checking or recovery?? (Score:3, Informative)
eric
The protocol implements in it's own way... (Score:2, Interesting)
As I see it, the real problems:
- SMP "experimentally" supported
- client and server can't coexist on same box
- client model is not decoupled enough from the server (a server going down can mean the client could crash)
It appears the driver software needs some work
Re:The protocol implements in it's own way... (Score:2)
- No client/server on the same box: Why the HELL would you want to do this? Most clients would have small local disks, if any, and mount most of their storage from a 'server' (possibly shared with other boxes using c.f. OpenGFS)
- Server crashing can crash a client: Most
SMP and why client/server on same box? (Score:2)
Since HyperSCSI is supposed to help you save money by enabling cheap hardware utility, I'd expect the flexibility to realize something like that to maximize the utility of connection-rich 1U/blade servers available today.
Also, don't you think it might be useful that a member of the disk array could access the
It's cute. but.... (Score:2, Insightful)
First, Ethernet can't be routed, so hyperSCSI isn't going to be nearly as flexible as iSCSI. This is the reason that just about everything that might want to be routed is usually carried over IP (and TCP and UDP and other stuff on top of IP). Straight ethernet is for stuff like ARP that really doesn't want to leave a network segment.
Of course, one could reasonably do something hyperSCSI-like across IP, and still save the TCP overhead. (Consider that in a low-loss short-hop environ
Re:It's cute. but.... (Score:2)
Yeah, it's five-nines reliablity, for a factory that makes 1,000,000 units a day, those 5 mistakes a day are gonna add up...
Re:It's cute. but.... (Score:2)
Lossy transport: HS implements its own flow-control.
The biggest concern i'd have is the lack of an integrity check. Most modern link layers do have their own their own integrity check, but usually pretty basic - they can miss errors. (see Linus' story on how his sources
sweet (Score:2)
Can multiple computers access the same drives through this?
On another note, is it possible to network over traditional SCSI, by changing the SCSI card ID's to make them co-operate on the same chain? Does an implementation of this exist in Linux, *BSD, or Solaris?
Re:sweet (Score:2)
All and all, a very interesting idea, even if it's not practicle (low number of devices on a SCSI bus, short cable length, etc). Maybe for a little mini-cluster of PC/104 boards if they had built in SCSI or something...
OK, after a quick googleing, I've found this [sourceforge.net] site. There are othe [google.com]
Re:sweet (Score:1)
I work for one of the first companies that is implementing Oracle RAC (2 and 4 node systems) over iSCSI to a Network Appliance.
I gotta tell you, it works great.
-djbkr
Single Vendor? (Score:1, Funny)
Didn't the article state that HyperSCSI is GPL and runs on Linux? What the fuck is this guy talking about?
Seriously, this is an excellent point (Score:2, Interesting)
The GPL establishes penalties for people or artificial entities that don't provide the sourcecode of the softwa
Comment removed (Score:3, Insightful)
NFS - history ignored (Score:5, Insightful)
The lessons of NFS are being ignored, and I'd expect HyperSCSI to die when it hits the same limitations.
NFS started out UDP-based, and moved toward TCP with NFSv3. Why? Because having all that error correction done at the network layer made for a better product; TCP does all the work to insure packets aren't lost or out-of-order. UDP doesn't, and the NFS application layer had to handle it, making it slower, more painful, and a duplication of effort better spent elsewhere.
The industry guys are almost right on this one. It isn't a beer can with a motor; it's a beer can with an M-80. Fun to watch when it works right, damn painful if you screw it up.
Re:NFS - history ignored (Score:5, Interesting)
It's worth adding, however, that the hyperSCSI folks are trying to make a distinction between wider-area networks (which they call SWANs for Storage WAN) and local single-segment (since they aren't routing) networks, and arguing that iSCSI is right for the former and hyperSCSI, because it's faster/cheaper, for the latter.
This view has parallels in the history of NFS over TCP versus NFS over UDP, because NFS/UDP is still hanging on in one niche: short-haul, high-speed, low-latency, few-hops, negligible-loss environments.
It also has parallels with the bad old days when direct-attached storage interconnects were much faster than LANs, so one set of protocols (FCP, SCSI, ESDI, IDE, SIMD...)evolved on the short fat pipes used to connect computers to peripherals, and a completely different set of protocols (ethernet, TCP/IP, SDLP,
Similarly, hyperSCSI is an argument that the two domains are different enough to justify different protocols. That seems to be arguing against a historical trend tht says that the short/fat and long/thin differences are vanishing; compare gigE and fibrechannel as _wires_ today.
All of this just reinforces Bourne's general point about ignoring the history. It's pretty clear that NFS over TCP is where the world is going, and the only reason that there's an NFS over UDP hanging around is that's how all NFS used to be, so some still is. When we compare hyperSCSI to iSCSI over TCP, I can't find any reason not to just deploy iSCSI everywhere and be done with it.
we need reliable connectionless messages (Score:2)
There are a number of choices. Plan 9 defined IL for just this purpose, there is SCTP for telephony applications, and several operating systems support RDM ("reliably delivered message"). Maybe people should just start using one of them, rather than reinventing the wheel by building on top of unrelia
Protocol/implementation confusion (Score:2)
You know it's coming soon... (Score:2)
Well, let's just add that to SCSI, SCSI 2, Fast SCSI, Wide SCSI, Fast Wide SCSI, Narrow SCSI, Ultra SCSI (aka SCSI 3), Ultra-2 SCSI, Ultra-3 SCSI (Ultra-160 SCSI to some), Ultra-320 SCSI and iSCSI. (I'm sure I've missed something out.)
So what's next for this party? UberSCSI? 1337SCSI? TheOneRingSCSI?
Re:You know it's coming soon... (Score:2)
One protocol to rule them all
One protocol to find them
One protocol to bring them all
And in the darkness to bind them.
So it's a master-slave-model distributed filesystem with autosensing of new nodes, that automatically propogates changes to the master node and binds additions to automagically generated mount points.
Neat.
Re:You know it's coming soon... (Score:2)
I thought BIND already had a patch to save us from the darkness [verisign.com]
Re:You know it's coming soon... (Score:2)
Implemented on raw token ring, of course.
Re:You know it's coming soon... (Score:2)
Re:You know it's coming soon... (Score:2)
NOOOOOOOOOO!!!! Tolken Ring revisited!
This looks interesting, however: (Score:4, Funny)
KFG
Give it a name... (Score:2)
Re:Give it a name... (Score:3, Funny)
Not to mention the potential cost overruns.
KFG
Re:This looks interesting, however: (Score:2)
iSCSI is a recognized standard (Score:2, Interesting)
iSCSI has drivers for every OS you can imagine, written by CISCO, IBM, Microsoft, and released under the GPL. This is from the iscsi sourceforge page.
This could get neat if used right.... (Score:2)
1: Create 2 networks, one being the normal network, and one being the SAN
2: Use GigE cable on your SAN with the inclusion of advanced Network FS'es like Coda.
3: Provide POP's that connect the external network with the internal SAN/serverNET by way of tunnels and port forwarding.
can you boot from it? (Score:3)
Biased article (Score:2, Troll)
"Without TCP/IP, it has no real error-recovery mechanism or guarantee that packets get delivered."
But that is wrong. There is error checking in the ethernet hardware and in the SCSI stack. It seems Smith needs to review the basic material, or should have at least read the introductory material [a-star.edu.sg]. Perhaps the takeaway here is, managers should not be allowed to comment on technical material, or if they do, they should solicit advice from a practicing engineer first.
Smith also dumps
Re:Biased article (Score:2)
Sorry, Daniel :) As Linux's network driver maintainer and author of a Serial ATA stack which goes through the Linux kernel SCSI layer... as well as being someone who reviewed ATA-over-ethernet [coraid.com] ...
I can say that ethernet hardware and the
Re:Biased article (Score:2)
I can say that ethernet hardware and the Linux SCSI stack does not handle retranmits that are needed at the ethernet layer. HP's Graham Smith is precisely correct. As some other slashdotters pointed out, the HyperSCSI code includes logic to hand
Re:Biased article (Score:2)
Sorry, Daniel
I can say that ethernet hardware and the Linux S
Re:why is there an article on this (Score:3, Funny)
Both of you are idiots, RTFA. (Score:3, Informative)
and
obviously, it's so when linux does support it, legions of slashbots can complain about the duplicate stories!
From the article:
Re:Enough of those double standards! (Score:2)
SCSI is dead? For the home computer perhaps, but not in the data center. I have yet to see a serial ATA enclosure perform on anything like the level of enclosures using SCSI over FC. Why don't you go look around a large office environment or a data center and see how many storage devices are SCSI based. This stuff (iSCSI, HyperSCSI, Ultra320, etc.) are aimed at servers and storage networks, not your desktop PC.
Re:Enough of those double standards! (Score:2)
SCSI may try to continue living, but it got only one direction to go: history's garbage pail.
You do realize that the serious data center hasn't run the SCSI protocol on a SCSI cable for years, right? We run FC everywhere. Even for our internal disks on some servers (Sun Vx80 series). Serial ATA doesn't even come close to SCSI over FC. My Hitachi 9570V gives me 19.5 TB raw storage and more than 100,000 IOPS. Let's see serial ATA and USB2 do that.
One more time, HyperSCSI, like FC and iSCSI is NOT for a de
I don't think hyperscsi is for the desktop either. (Score:2)
I can see small offices, workgroups, and studios using this to get high-speed storage on line quickly with cheap components.
iSCSI has a better chance of being deployed in the home (of Unix-y types and their unsuspecting kin). I'd say, iSCSI all the way, none of this baby CIFS BS in my ho
Re:Enough of those double standards! (Score:3, Funny)
Many people here disagree with you. I wish I had SCSI hardware... you troll... now you've hurt my feelings...
Re:Enough of those double standards! (Score:1)
Re:Enough of those double standards! (Score:5, Insightful)
More to the point, we have usb 2, serial ata, and firewire. Really all you should need is usb 1.1 and firewire in ever-increasing speeds, besides your display output. There's literally no need for anything else. Firewire can be operated in a synchronous fashion, after all. It supports lots of devices per adapter (63 or 127 depending on implementation) and comes in powered and unpowered as well as peer to peer or host-based forms, but nearly all devices will fall back to the lowest, slowest mode - which is capable of 400Mbps per second, or 50 Megabytes, clearly enough for all but the fastest hard drives, and nearly any other use to which you might put it. 800Mbps firewire is available today, and 1.6Gbps is uh, just over the horizon or something. They claim they'll do 3.2Gbps over fiber in the near future as well, but I'll just stick to developments just over the horizon, not way over it, here.
Frankly what I want to see more than anything is native 1394 storage devices. 1394 is somewhat to scsi (it would kind of have to be, wouldn't it?) and is at least well documented. I want hard drives that have only power and 6 pin 1394 connections on them, pass through style. You could treat them basically like an external scsi chain not requiring termination, but inside your PC. Firewire to IDE bridges are not the answer, they just make things more complicated and firewire is supposed to make things less complicated, though I suppose they are a sort of interim solution.
For the average user, if you could get native (and thus inexpensive) firewire solutions for everything, it would fulfill your needs much better and would make computer use much simpler. The only thing you need besides firewire and USB for every common peripheral is video output and possibly high speed networking, perhaps provided from a MII or similar connection, just as AUI was the standard at one time. Just, without the annoying box and cables. And let us not forget, a memory slot. I assume most people would also like to have wireless ethernet with an external antenna jack. For the rest of us, we would continue to run our however-shaped boxes with internal expansion, PCI-Express or what have you, and our cabling would be much simplified. Once again, 800Mbps is 100MBps, yes? Assuming you can only really sustain 50 or 60% of that with multiple devices competing for the bus, that's still fast enough for basically any two hard drives anyone is likely to have at home. So 800Mbps firewire is perfectly adequate to the task of connecting hard drives today. It's not that SATA isn't, it's that having less types of interface simplify a computer. If you're only going to have two interfaces, you will get more mileage out of USB2 and IEEE1394 than you will USB2 and SATA, even one of the bastardized forms of it like External SATA, if people will just make IEEE1394 devices.
Re:Enough of those double standards! (Score:2)
Absolutely right, except for teh monitor... Firewire actually IS fast enough to supply data to your monitor.
But even better, is the fact that you could have an unlimited number of drives, either inside or outside of your case (it really wouldn't matt
Re:Enough of those double standards! (Score:2)
Re:Enough of those double standards! (Score:2)
Yeah, and then there's the fact that it'll be a cold day in hell when 'they' start building their primary IO around an Apple-invented technology.
Ignoring for now the similarities between AppleTalk and USB, NuBus and plug-n-play PCI, integrated sound chips on every motherboard, etc. Gee, looks like they'll have to come up with SerialATA version
Re:Enough of those double standards! (Score:5, Informative)
Serial ATA maximum cable length: 1m
Fiber Channel maximum cable length: 10,000 m
Re:Enough of those double standards! (Score:4, Interesting)
Fiber Channel maximum cable length: 10,000 m
Add the appropriate routers and switches and you can easily go 90 km on dark fiber. Add some appropriate routers onto a fast network (T3, ATM, what have you) and you can go 500 km. With fast FC connected storage at each end. Of course, this sort of solution is used by data centers, not home users. But Fiber is the obvious solution to data storage problems. And there is enough mass in the server storage market now that prices are starting to come down. Of course, if you need fast, redundant, capable storage you won't blink at the cost.
Re:Enough of those double standards! (Score:2)
It's not going to be dark fibre if you're using it, is it??
Re:Enough of those double standards! (Score:2)
Re:Enough of those double standards! (Score:2)
So you can keep your pr0n collection in your secret undergound bunker, along with your spare deflector beanies.
"If I want peripherals to communicate for such a distance, I'll just use LAN."
A LAN? That reaches ten klicks? Are we not clear on what the "L" in "LAN" stands for?
Re:Enough of those double standards! (Score:2)
Just because you cannot fathom a use for a particular technology does not mean that a technology is useless. Some companies find the ability to perform block I/O over long distances to be very useful.
Re:Enough of those double standards! (Score:2, Informative)
Come back and troll when SATA has doubled in speed, and when I can plug at least 15 drives into a card.
Or hell, just stay out of the game, since Fibre Channel has 2Gbit/ps (250MB/ps, still faster than SATA, slower than ultra320) and 255 devices, with multiple host access over a SAN, which can be set up redundantly. And that ignores the point of this article, S
Re:Enough of those double standards! (Score:2, Informative)
Re:That nasty TCP/IP overhead (Score:4, Informative)
IP checksum and TCP checksum are necessary (Score:5, Interesting)
Layer Total packets # chksum errs
Ethernet 170,000,000 446
IP 170,000,000 14
UDP 140,000,000 5
TCP 30,000,000 350
Basically, when absolute accuracy is required the more error checking the better.
Re:Ethernet is NOT the right media (Score:2)
Re:Ethernet is NOT the right media (Score:2)
As long as our OS'es run on multiple platforms with good hardware behind them, Intel's just a low end cheap solution. After all, the better chipsets are using the Gx series (care of Motorola and IBM) or Sparc hardware.
Intel, AMD and Cyrix have enabled cheap, low end mainframe computing for the masses... After all, look at some of the better OS'es - QNX, Linux, FreeBSD, Solaris.
Re:concurrent filesystem access (Score:3)
In short, it doesn't. That has to be done by the filesystem layer or application layer itself. A Fibre Channel-based SAN setup such as is common in enterprise deployments doesn't normally provide for this either (Fibre Channel zoning is an exception, but that's a feature not normally required by most setups). If you really want to do concurrent filesystem access by multiple machines, you need a distributed lock manager of some kind, similar to what you have in Oracle RAC's cluster filesystem, Sistina's G [sistina.com]