In scientists' quest to understand the true nature of the universe and grapple with the mysterious dark matter and dark energy that seemingly accounts for 95 percent of the universe, they have built ever-more powerful and ingenious ways of studying the skies.
Over the coming years, this will allow researchers to construct simulations of the universe at previously inconceivable levels of detail. But to do so, those researchers will need mind-bendingly powerful computers and algorithms capable of extreme scaling. Now a US team of astrophysicists have published their plans to create the most detailed universe simulation ever.
The so-called Hybrid/Hardware Accelerated Cosmology Code (Hacc) provides a novel framework for cosmological simulation, which the team, comprising researchers from the Argonne, Los Almos and Lawrence Berkley US national laboratories, have shown is capable of generating simulations with 3.6 trillion particles. That is “significantly bigger than any cosmological simulation yet performed,” the team claimed.
To demonstrate the validity of their Hacc framework, the team grabbed some time on one of IBM's third generation of BlueGene supercomputers, the BG/Q. A single BG/Q rack contains 1,024 nodes, each with 16GB of DDR3 memory, and a BG/Q compute chip which uses 17 augmented 64-bit PowerPC A2 cores.
They were able to show that Hacc achieved massive scalable performance – 13.94 Pflops/s (quadrillions of calculations per second) at 69.2 percent of peak and 90 percent parallel efﬁciency on 1,572,864 cores.
Those are big numbers, but to give it some comparison, the IBM BlueGene Sequoia at the Lawrence Livermore National Laboratory was until recently ranked as the most powerful supercomputer in the world. In June 2012, it posted a Linpack benchmark of 16.32 Pflops/s.
Still, with that sort of power, the team are confident that Hacc will be capable of creating detailed simulations that could use measurements of weak gravitational lensing to map the distribution of dark matter throughout the universe, or of the distribution of galaxies and clusters, from the largest to the smallest scales.
The team are now overseeing the acceptance tests necessary to get Hacc operating on Livermore's Sequoia as well as on the Mira supercomputer at the Argonne National Laboratory.
The team's work was presented at the SC12 conference in Salt Lake City, Utah this week.
IBM and Technical University of Munich team demonstrate how Shor's algorithm, which can't be cracked by conventional computers, can be solved quickly with quantum computing
Hubble Space Telescope finds superflares from young red dwarfs could strip away planetary atmosphere
Younger stars are 100 to 1,000 times more energetic than when they're older
Two of the big four supermarkets will use the system to control sales of restricted products
PUBG news and updates: November's Update #23 to bring new Skorpion pistol and changes to blue zone visibility
Genuinely useful side-arm coming to PUBG in Update #23