OSC storing vast data from European supercollider

Using the world’s most powerful particle accelerator and mass data storage sites, such as the Ohio Supercomputer Center, more than 1,000 international physicists, engineers and technicians – including Ohio State University Professor Thomas J. Humanic – now have begun the process to unravel questions about the first moments of the universe.

In April, physicists working on the ALICE project (short for A Large Ion Collider Experiment) began recording data from collisions within the Large Hadron Collider, operated by the European Laboratory for Nuclear Research (CERN) near Geneva, Switzerland. The researchers hope to find answers to fundamental questions about the birth of the universe, matter vs. antimatter, the nature of dark matter and maybe even the existence of other dimensions.

The ALICE scientists employ a series of nearly 9,600 powerful magnets to carefully propel and collide opposing beams of protons, and beams of lead nuclei later this year, at nearly the speed of light around a 17-mile underground loop. The proton-proton collisions were conducted at seven tera electron-volts (TeV, a unit of momentum in high energy physics). These are the highest energy proton collisions ever produced in the laboratory – 3.5 times higher than the previous highest energy proton collisions created at the Tevatron particle collider, located near Chicago at the Department of Energy’s Fermilab.

The ALICE collisions expel hundreds to thousands of small particles, including quarks – which make up the protons and neutrons of the atomic nuclei – and gluons – which bind the quarks together. For a fraction of a second, these particles form a fiery-hot plasma that hasn’t existed since the first moments after the Big Bang, about 14 billion years ago. Within the massive 52-foot ALICE detector, 18 sensitive sub-detectors measure the behavior of the expelled particles, recording up to approximately 1.25 gigabytes of data per second – six times the contents of Encyclopedia Britannica every second.

The massive data sets are now being collected and distributed to researchers around the world through high-speed connections to the LHC Computing Grid (LCG), a network of computer clusters at scientific institutions, including the Ohio Supercomputer Center. The network employs the connectivity of private fiber-optic cable links, as well as existing portions of the public Internet.

The LCG is composed of more than 100,000 processors at 130 organizations across 34 countries and is organized into four levels, or ‘tiers.’ Tier 0 is CERN’s central computer, which distributes data to the eleven Tier 1 sites around the world. The Tier 1 sites, in turn, coordinate and send data to Tier 2 sites, which are centers that provide storage capacity and computational analysis for specific tasks. Scientists access the stored data through Tier 3 sites – individual computers operated at research facilities.

“Traditionally, researchers would do much, if not all, of their computing at one central computing center. This cannot be done with the ALICE experiments because of the large data volumes,” said Humanic. “OSC has been contributing computing resources to the project from the very beginning of ALICE’s distributed computing efforts, starting in 2000.”

Construction of the LHC began in 1995, when much of the necessary computational and networking technologies didn’t yet exist. The long-term plan for the project loosely relied upon a concept referred to as Moore’s law, which describes the trend of computer processing power doubling every two years.

“If CERN had brought the LHC online much sooner, computing centers would have had a problem meeting the challenges,” said Doug Johnson, a senior systems developer at OSC. “For quite some time, OSC has been moving to meet the needs of a different mode of research, where computers analyze the huge amounts of otherwise raw data collected from instruments, such as satellites, microscopes, sequencing machines and particle colliders.”

A prime example of the development of new technologies to meet the demands of the ALICE project is the emergence of grid computing technologies. “OSC was one of the first adopters of the ALICE-developed AliEn (ALICE Environment) grid infrastructure,” Humanic explained.

As a Tier-2 site on the LCG, OSC this year has committed, through its normal allocations process, 30 terabytes of data storage and one million processor hours – equal to about 115 home computers running 24 hours a day for a year, according to Johnson.

“This data will be accessed by Dr. Humanic and his OSU colleagues, as well as researchers anywhere in the world, for downloading, reconstruction and analysis,” Johnson said. “Researchers can sit at their laptops, write small programs or macros, submit the programs through the AliEn system, find the necessary ALICE data on AliEn servers and then run their jobs through centers such as OSC.”

Beyond serving as a storage and analysis resource for researchers working on the project, “OSC also has been critical in the development and testing of a computing model to analyze the ALICE data,” Humanic said. OSC had provided 300,000 CPU hours for data simulations prior to actual experiments at the LHC.

Ohio State is the only institution in the U.S. collaborating on three of the LHC’s four major detectors. Beyond ALICE, researchers within the OSU Department of Physics also are collaborating on two other large-scale LHC experiments: the ATLAS (A Toroidal LHC Apparatus) and CMS (Compact Muon Solenoid) projects. Each of the three projects uses general-purpose detectors to analyze particles caused by collisions produced by the accelerator. The OSU-OSC efforts are jointly funded by the National Science Foundation and the Department of Energy.