The Top500 list is compiled twice-yearly by a team led by a professor from Germany's University of Mannheim.

It measures how fast the computers can solve a special type of linear equation to determine their speed, but does not take account of other factors - such as how fast data can be transferred from one part of the system to another - which can also influence real-world performance.

Fastest supercomputers

1. Tianhe-2 (China) 33.86 petaflop/sec

2. Titan (US) 17.59 petaflop/sec

3. Sequoia (US) 17.17 petaflop/sec

4. K computer (Japan) 10.51 petaflop/sec

5. Mira (US) 8.59 petaflop/sec

6. Piz Daint (Swiss) 6.27 petaflop/sec

7. Stampede (US) 5.17 petaflop/sec

8. Juqueen (Germany) 5.09 petaflop/sec

9. Vulcan (US) 4.29 petaflop/sec

10. SuperMuc (Germany) 2.90 petaflop/sec

(Source: Top500 List based on Rmax Linpack benchmark)

IBM - which created five out of the 10 fastest supercomputers in the latest list - told the BBC it believed the way the list was calculated should now be updated, and would press for the change at a conference being held this week in Denver, Colorado.

"The Top500 has been a very useful tool in the past decades to try to have a single number that could be used to measure the performance and the evolution of high-performance computing," said Dr Alessandro Curioni, head of the computational sciences department at IBM's Zurich research lab.

"[But] today we need a more practical measurement that reflects the real use of these supercomputers based on their most important applications.

"We use supercomputers to solve real problems - to push science forward, to help innovation, and ultimately to make our lives better.

"So, one thing that myself and some of my colleagues will do is discuss with the Top500 organisers adding in new measurements."

However, one of the list's creators suggested the request would be denied.

"A very simple benchmark, like the Linpack, cannot reflect the reality of how many real application perform on today's complex computer systems," said Erich Strohmaier.

"More representative benchmarks have to be much more complex in their coding, their execution and how many aspects of their performance need to be recorded and published. This makes understanding their behaviour more difficult.

"Finding a good middle-ground between these extremes has proven to be very difficult, as unfortunately all previous attempts found critics from both camps and were not widely adopted."

China's lead

Tianhe-2 - which translates as Milky Way 2 - was developed by China's National University of Defence Technology and will be based in the city of Guangzhou, in the country's south-eastern Guandong province.

Image caption
IBM's Sequoia recently carried out a simulation of a cloud of 15,000 bubbles

It uses a mixture of processors made by Intel as well as custom-made CPUs (central processing units) designed by the university itself.

The system is to be offered as a "research and education" tool once tests are completed, with local reports suggesting that officials have picked the car industry as a "priority" client.

Its Linpack score is nearly double that of the next supercomputer in the list - Titan, the US Department of Energy's system at the Oak Ridge National Laboratory in Tennessee.

However, one expert said it was still too early to know whether the Chinese system would be able to outperform its US counterpart in real-world tasks.

"You can get bottlenecks," said Prof Alan Woodward, from the University of Surrey's department of computing.

"Talking about the number of calculations that can be carried out per second isn't the same as saying a supercomputer can do that in practice in a sustained way. The processors might be kicking their heels some of the time if they don't get the data as fast as they can handle, for example."

Energy efficiency

Supercomputer applications do not tend to use all the processor power on offer.

IBM notes that its own Sequoia supercomputer - which came third on the latest list - used a relatively high 73% of the machine's theoretical peak performance when it recently carried out what the firm describes as the biggest ever fluid dynamics simulation to date.

Image caption
Switzerland's Piz Daint will be used to model galaxy formations and weather patterns