Speeding up extreme big brain data analysis

It's tough to unravel the mysteries of the brain when your computer is frozen.

To aid frustrated brain researchers, a multidisciplinary team of scientists at the University of Utah has created a faster method for generating and exploring high-resolution, 3-D images of the brain.

Alessandra Angelucci, a neuroscientist with the Moran Eye Center at Utah's Ophthalmology Department whose research is funded by the National Science Foundation (NSF), says the tool allows her and her colleagues to better visualize the complex connections between neurons.

"We know that brain function emerges from the activity of neural networks," Angelucci says. "To understand how the brain works, we need to know how it is wired."

The work by the Utah team is part of a movement across the scientific community to come up with new ways to use and analyze information. Scientific instruments ranging from brain scanners, to microscopes, to telescopes are increasingly improving and generating larger amounts of high-quality data, creating the need for more efficient data-processing approaches.

"This is really an extreme big data project," says Cameron Christensen, a Utah computer scientist on a team led by Valerio Pascucci at the SCI Institute at Utah, who created the new method at the Center for Extreme Data Management Analysis and Visualization. "We're living in a time of extreme big data. This tool could have applications for astronomy, climate change and other fields."

Big data tools are vital to advance our understanding of the brain, universe and other topics, as evidenced by national efforts such as The BRAIN Initiative.

New method

The tool is a new twist on an existing technology.

The technology, known as Visualization Streams for Ultimate Scalability (ViSUS), turns massive quantities of digital data into visual information that the human mind can understand. ViSUS has been used for years in several different areas of research, including clean energy simulations on high performance computing platforms.

The software processes terascale-sized datasets in a way that allows researchers to explore the images interactively. The typical laptop has about 4-15 gigabytes of memory. One terabyte has 1,000 times that much. That's doable for a supercomputer, but not a grad student's Mac.

"Previously, if you wanted to interact with this volume of data, you couldn't. Everything would freeze," Christensen says.

Think of the desk-smacking frustration involved in trying to open an image file 1,000 times greater than your computer can handle.

This is a frustration brain researchers know well.

A whole mouse brain is about 31 terabytes of imaging data. That is same amount of data as about 32 million digital photographs.

The visual cortex of the macaque brain, which Angelucci studies, is 318 terabytes.

"We call these 'terror-bytes,'" said Terry Sejnowski of the Salk Institute for Biological Sciences, a computational neuroscientist who moderated the Society for Neuroscience press conference at which the findings were announced.

The new approach developed by Pascucci's team to handle these terror-bytes is hierarchical, loading images of increasing resolution as the user clicks to see more details.

"It's a little like Google Maps, where you zoom in and in. Except with our technique, you don't reload all the data each time. You only download the differences between levels, so it's faster and requires less memory," he says. "Therefore, it opens the possibility of dealing with 3-D microscopy volumes instead of just 2-D maps."

Millimeters of brain

Angelucci and her colleagues used the technology to stream 3-D images of a primate's visual cortex.

An NSF-funded team has developed a tool that allows brain researchers to better visualize the complex connections between neurons. Shown here is a zoomed in region of the visual cortex in a macaque brain. Neuroscientists can more easily navigate blood vessels (orange) and neurons (green).

Credit: Cameron Christensen, University of Utah

Angelucci and her colleagues used the technology to stream 3-D images of a primate's visual cortex. They paired this tool with a neuroscience technique called CLARITY, which makes surrounding brain tissue transparent. (CLARITY was developed by NSF-funded researchers in 2013.) Shown here is a brain becoming transparent, thanks to CLARITY.

Credit: Frederick Federer, University of Utah

The technology, known as Visualization Streams for Ultimate Scalability (ViSUS), has been used for years in several different areas of research, including clean energy simulations on High Performance Computing platforms. The software processes interactively terascale-sized datasets. The typical laptop has about 4-to-15 gigabytes of memory. One terabyte has 1,000 times that much. That's doable for a supercomputer, but not a grad student's Mac. A whole mouse brain is about 31 terabytes of imaging data. That is same amount of data as about 32 million digital photographs. The visual cortex of the macaque brain, which Angelucci studies, is 318 terabytes -- 100 times more.

Credit: Frederick Federer, University of Utah

There is great need for more efficient data-processing approaches as microscopes, telescopes and other instruments improve to acquire larger amounts of high-quality data. Big data tools are vital to advance our understanding of the brain, universe and other topics, as evidenced by national efforts such as The BRAIN Initiative. Shown here is a portion of the macaque visual cortex shown using the ViSUS Powerwall viewer.

Credit: Vincent Horiuchi, University of Utah

Data can be explored interactively using a variety of visualization methods. Shown here is high-resolution image of a macaque visual cortex. The neurons are in a 2.5 millimeter-squared region.

The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. In fiscal year (FY) 2015, its budget is $7.3 billion. NSF funds reach all 50 states through grants to nearly 2,000 colleges, universities and other institutions. Each year, NSF receives about 48,000 competitive proposals for funding, and makes about 11,000 new funding awards. NSF also awards about $626 million in professional and service contracts yearly.