Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×

Storing CERN's Search for God (Particles) 154

Chris Lindquist writes "Think your storage headaches are big? When it goes live in 2008, CERN's ALICE experiment will use 500 optical fiber links to feed particle collision data to hundreds of PCs at a rate of 1GB/second, every second, for a month. 'During this one month, we need a huge disk buffer,' says Pierre Vande Vyvre, CERN's project leader for data acquisition. One might call that an understatement. CIO.com's story has more details about the project and the SAN tasked with catching the flood of data."
This discussion has been archived. No new comments can be posted.

Storing CERN's Search for God (Particles)

Comments Filter:
  • by xxxJonBoyxxx ( 565205 ) on Saturday July 21, 2007 @12:34AM (#19935629)
    Um...no. Actually, it's a product placement PR piece about Quantum's StorNext. (Read page 2...)
  • by snowraver1 ( 1052510 ) on Saturday July 21, 2007 @12:55AM (#19935717)
    2.6 Petabytes. The article says that they will be collecting petabytes of data. Also, the article clearly said GB. GB= Gigabyte Gb= Gigabit. The thing that I thought was "Wow that's ALOT of blinking lights!" Sweet!
  • Re:PC's? (Score:2, Informative)

    by Falstius ( 963333 ) on Saturday July 21, 2007 @12:59AM (#19935731)
    Actually, there really is a gigantic room at CERN full of commodity PCs that form the first level of computing for the different experiments. The data is then shipped off to sites around the world for further processing. There is a combination of 'locally' distributed computing and world-wide grid being used.
  • Re:PC's? (Score:1, Informative)

    by Anonymous Coward on Saturday July 21, 2007 @01:56AM (#19935909)
    Initially some data is being filtered at the detector pits by the farms of PCs doing the triggering. After that the data will be fed to storage and analysis. CERN has been upgrading its computer centre for quite some while (the main problem is not power supply, but cooling system - thus some of performance benchmarks also include it). Besides CERN (Tier-0) will have high-speed connections (via means of LCG backbone) with many sites around the world and the data processing will be done in a 'global manner'.

    You can google on phrase 'service challenge' site:cern.ch, or just go to the LCG [web.cern.ch] site.

    --
    Milosz [hulboj.org]
  • Re:News for Nerds! (Score:5, Informative)

    by xyvimur ( 268026 ) <koo3ahzi AT hulboj DOT org> on Saturday July 21, 2007 @02:10AM (#19935971) Homepage
    Just go there and take a guided tour. If you'll hurry you'll be able to go to the detector pit and see it. Otherwise after starting up it will be inaccesible for visitors for the life-cycle of the experiments (10-20 years). Google for CERN visit service.

    Milosz [hulboj.org]
  • by noggin143 ( 870087 ) on Saturday July 21, 2007 @02:42AM (#19936065)
    We are expecting to record around 15PB / year during the LHC running. This data is stored onto magnetic tape with petabytes of disk cache to give reasonable performance. A grid of machines distributed worldwide analyses the data. More details are available on the CERN web site www.cern.ch.
  • Re:PC's? (Score:5, Informative)

    by Rodolpho Zatanas ( 986694 ) on Saturday July 21, 2007 @02:42AM (#19936067) Homepage
    load"*",8,1 would load something from a diskette, not a cassette.
  • Not So Huge (Score:5, Informative)

    by PenGun ( 794213 ) on Saturday July 21, 2007 @02:45AM (#19936083) Homepage
    It's only 5x HD SDI single channel ~ 200MB/s. Any major studio could handle this with ease.

    SDI is how the movie guys move their digital stuff around. A higher end digital camera will capture at 2x HD SDI for a 2K res, 4:4:4 colour space. A few of em' and you got your 1GB/s easy. Spools onto godlike RAID arrays.

      Get em' to call up Warner Bros if they have problems.
  • by this great guy ( 922511 ) on Saturday July 21, 2007 @02:46AM (#19936089)

    Assuming a non-RAID 3x-replication tech solution (what Google do in their datacenters), using 500-GB disks (best $/GB ratio), they would need about 16 thousands disks:

    .001 (TB/sec) * 3600*24*30 (sec/month) * 3 (copies) * 2 (disk/TB) = 15552 disks

    Which would cost about $1.8M (disks alone):

    15552 (disk) * 110 ($/disk) = $1710720

    Packed in high-density chassis (48 disks in 4U, or 12 disks per rack unit), they could store this amount of data in about 30 racks:

    15552 (disk) / 12 (disk/rack unit) / 42 (rack unit/rack) = 30.9 racks

    Now for various reasons (vendors influence, inexperienced consultants, my experience in the IT world in general, etc), I have a feeling they are going to end up with a solution unnecessarily complex, much more expensive, and hard to maintain and expand... Damn, I would love to be this project leader !

  • by Roger W Moore ( 538166 ) on Saturday July 21, 2007 @04:48AM (#19936423) Journal
    The ALICE experiment is actually concentrating on heavy ion collisions which is why they only worry mainly about one month/year, the rest of the time the machine is running protons for the other experiments, ATLAS and CMS, which will look for the Higgs. ALICE will hopefully study the quark gluon plasma but, as far as I know, has no plans to look for the Higgs.
  • by quarkie68 ( 1018634 ) on Saturday July 21, 2007 @06:12AM (#19936747) Homepage

    OK, we got a half way overview of CERN's decision, with some bold statements of questionable validity. I am submitting the criticism purely on the grounds of being really interested in large data storage, I don't work for any large storage vendor, but I am an architect of storage systems.

    First of all, with the statement "and it's (StorNext) completely vendor independent": Lot's of other solutions provide flexibility about choosing the hardware vendor from a theoretical perspective. The theory says that if vendor A makes a SAN, vendor B makes a RAID controller, C a disk cabinet and D offers a clustered FS, and all comply to the relevant standards, you can plug them together and expect them to function. However, imperfections in the standards, hidden proprietary optimizations, always dictate certain configs and combinations for optimum performance. There is a lot of work to be done in the StorNext and other similar products, until they claim full flexibility. My experience in deploying a StorNext based solution on a 1200 node setup says so and to keep the post short, I shall exclude at this stage vendor details, but if someone is interested, I am happy to go over the details. There is vendor dependence if you wish optimum performance. Not to mention that if you mix and match the RAID and SAN cards in the setup, any unfortunate issue might end up in a multi-headache, even if you have solution support (A blaims B, B accusses A, and the game of ping-pong begins). You can never exclude vendor dependence in such a large setup, you have to deal with it.

    Then you have the "Clustered file systems are still an evolving category, she says, but enterprise IT is warming up to it.". I can imagine what the author classes as enterprise IT here, but I think there is a bit of an orientation issue. CERN is not exactly the classical enterprise IT environment, is it? Not in terms of their requirements for resilience and capacity. These FAR EXCEED enteprise IT requirements. CERN is a research setup. And the mentality of a research setup (that incubated the WWW after all) is (or should be) that of innovation and playing with some of the latest and the greatest. In fact, some US based research setups have long experimented with other cluster FSes. They are not warming up. CIO claims that StorNext is scalable. It is. But to what extent? Have they excluded for example things such as Lustre? http://wiki.lustre.org/index.php?title=Main_Page [lustre.org] If yes, why?

  • by Main Gauche ( 881147 ) on Saturday July 21, 2007 @12:10PM (#19938537)

    "Imagine how deep the personality problems must run in a person who gets all hot because of someone's DNA sequences!"

    You must be new here.

  • Try 5,000 years ago (Score:3, Informative)

    by benhocking ( 724439 ) <benjaminhocking@[ ]oo.com ['yah' in gap]> on Saturday July 21, 2007 @02:30PM (#19939569) Homepage Journal
    I think you're thinking of that guy who got nailed to the cross (Jesus). Noah was born about 5,000 [itgo.com] years ago.

Never test for an error condition you don't know how to handle. -- Steinbach

Working...