New network design exploits cheap, power-efficient flash memory without sacrificing speed

newnetworkde

Random-access memory, or RAM, is where computers like to store the data they’re working on. A processor can retrieve data from RAM tens of thousands of times more rapidly than it can from the computer’s disk drive.

But in the age of big data, data sets are often much too large to fit in a single computer’s RAM. The data describing a single human genome would take up the RAM of somewhere between 40 and 100 typical computers.

Flash memory—the type of memory used by most portable devices—could provide an alternative to conventional RAM for big-data applications. It’s about a tenth as expensive, and it consumes about a tenth as much power.
The problem is that it’s also a tenth as fast. But at the International Symposium on Computer Architecture in June, MIT researchers presented a new system that, for several common big-data applications, should make servers using flash memory as efficient as those using conventional RAM, while preserving their power and cost savings.
The researchers also presented experimental evidence showing that, if the servers executing a distributed computation have to go to disk for data even 5 percent of the time, their performance falls to a level that’s comparable with flash, anyway.
In other words, even without the researchers’ new techniques for accelerating data retrieval from flash memory, 40 servers with 10 terabytes’ worth of RAM couldn’t handle a 10.5-terabyte computation any better than 20 servers with 20 terabytes’ worth of flash memory, which would consume only a fraction as much power.
“This is not a replacement for DRAM [dynamic RAM] or anything like that,” says Arvind, the Johnson Professor of Computer Science and Engineering at MIT, whose group performed the new work. “But there may be many applications that can take advantage of this new style of architecture. Which companies recognize: Everybody’s experimenting with different aspects of flash. We’re just trying to establish another point in the design space.”
Joining Arvind on the new paper are Sang Woo Jun and Ming Liu, MIT graduate students in computer science and engineering and joint first authors; their fellow grad student Shuotao Xu; Sungjin Lee, a postdoc in Arvind’s group; Myron King and Jamey Hicks, who did their PhDs with Arvind and were researchers at Quanta Computer when the new system was developed; and one of their colleagues from Quanta, John Ankcorn—who is also an MIT alumnus.

Outsourced computation

The researchers were able to make a network of flash-based servers competitive with a network of RAM-based servers by moving a little computational power off of the servers and onto the chips that control the flash drives. By preprocessing some of the data on the flash drives before passing it back to the servers, those chips can make distributed computation much more efficient. And since the preprocessing algorithms are wired into the chips, they dispense with the computational overhead associated with running an operating system, maintaining a file system, and the like.
With hardware contributed by some of their sponsors—Quanta, Samsung, and Xilinx—the researchers built a prototype network of 20 servers. Each server was connected to a field-programmable gate array, or FPGA, a kind of chip that can be reprogrammed to mimic different types of electrical circuits. Each FPGA, in turn, was connected to two half-terabyte—or 500-gigabyte—flash chips and to the two FPGAs nearest it in the server rack.
Because the FPGAs were connected to each other, they created a very fast network that allowed any server to retrieve data from any flash drive. They also controlled the flash drives, which is no simple task: The controllers that come with modern commercial flash drives have as many as eight different processors and a gigabyte of working memory.
Finally, the FPGAs also executed the algorithms that preprocessed the data stored on the flash drives. The researchers tested three such algorithms, geared to three popular big-data applications. One is image search, or trying to find matches for a sample image in a huge database. Another is an implementation of Google’s PageRank algorithm, which assesses the importance of different Web pages that meet the same search criteria. And the third is an application called Memcached, which big, database-driven websites use to store frequently accessed information.

Chameleon clusters

FPGAs are about one-tenth as fast as purpose-built chips with hardwired circuits, but they’re much faster than central processing units using software to perform the same computations. Ordinarily, either they’re used to prototype new designs, or they’re used in niche products whose sales volumes are too small to warrant the high cost of manufacturing purpose-built chips.
But the MIT and Quanta researchers’ design suggests a new use for FPGAs: A host of applications could benefit from accelerators like the three the researchers designed. And since FPGAs are reprogrammable, they could be loaded with different accelerators, depending on the application. That could lead to distributed processing systems that lose little versatility while providing major savings in energy and cost.
“Many big-data applications require real-time or fast responses,” says Jihong Kim, a professor of computer science and engineering at Seoul National University. “For such applications, BlueDBM”—the MIT and Quanta researchers’ system—”is an appealing solution.”
Relative to some other proposals for streamlining big-data analysis, “The main advantage of BlueDBM might be that it can easily scale up to a lot bigger storage system with specialized accelerated supports,” Kim says.
References:http://phys.org/

Computer program fixes old code faster than expert engineers

computerprog

“The order of operations in these optimized binaries are complicated, which means that they can be hard to disentangle,” says Mendis, a graduate student at CSAIL. “Because stencils do the same computation over and over again, we are able to accumulate enough data to recover the original algorithms.”

From there, the Helium system then replaces the original bit-rotted components with the re-optimized ones. The net result: Helium can improve the performance of certain Photoshop filters by 75 percent, and the performance of less optimized programs such as Microsoft Windows’ IrfanView by 400 to 500 percent.

“We’ve found that Helium can make updates in one day that would take human engineers upwards of three months,” says Amarasinghe. “A system like this can help companies make sure that the next generation of code is faster, and save them the trouble of putting 100 people on these sorts of problems.”

The research was presented in a paper accepted to the Association for Computing Machinery SIGPLAN conference on Programming Language Design and Implementation (PLDI 2015), which took place June 13-17 in Portland, Oregon.

The paper was written by Mendis, fellow graduate students Jeffrey Bosboom and Kevin Wu, research scientist Shoaib Kamil, postdoc Jonathan Ragan-Kelley PhD ’14, Amarasinghe, and researchers from Adobe and Google.

“We are in an era where computer architectures are changing at a dramatic rate, which makes it important to write code that can work on multiple platforms,” says Mary Hall, a professor at the University of Utah’s School of Computing. “Helium is an interesting approach that has the potential to facilitate higher-level descriptions of stencil computations that could then be more easily ported to future architectures.”

One unexpected byproduct of the work is that it lets researchers see the different tricks that programmers used on the old code, such as archaeologists combing through computational fossils.

“We can see the ‘bit hacks’ that engineers use to optimize their algorithms,” says Amarasinghe, “as well as better understand the larger context of how programmers approach different coding challenges.”

References:http://phys.org/

Samsung Galaxy S6 Active vs. iPhone 6

iphone-6-vs-galaxy-s6-active

Gizmag compares the Samsung Galaxy S6 Active (left) to the Apple iPhone 6

For the third straight year, Samsung is giving you the option of buying a rugged version of its Galaxy flagship. Let’s compare the features and specs of the Galaxy S6 Active to the iPhone 6.

iphone-6-vs-galaxy-s6-active-8

New energy cell can store up solar energy for release at night

renewable-energy-hybrid-storage-cell

Researchers have, for the first time, found a way to store electrons generated by photoelectrochemical (PEC) cells for long periods of time

A photoelectrochemical cell (PEC) is a special type of solar cell that gathers the Sun’s energy and transforms it into either electricity or chemical energy used to split water and produce hydrogen for use in fuel cells. In an advance that could help this clean energy source play a stronger role within the smart grid, researchers at the University of Texas, Arlington have found a way to store the electricity generated by a PEC cell for extended periods of time and allow electricity to be delivered around the clock.

Currently, the electricity generated by a PEC cell could not be stored effectively, as the electrons would quickly “disappear” into a lower-energy state. This meant that these cells were not a viable solution for a clean-energy grid, as the electricity had to be used very shortly after being produced. That is, on sunny days, at a time when standard PV panels would already be producing energy at full tilt.

Now, researchers Fuqiang Liu and colleagues have created a PEC cell that includes a specially designed photoelectrode (the component that converts incoming photons into electrons). Unlike previous designs, their hybrid tungsten trioxide/titanium dioxide (WO3/TiO2) photoelectrode can store electrons effectively for long periods of time, paving the way for PEC cells to play a bigger role within a smart energy grid.

renewable-energy-hybrid-storage-cell-2

The system also includes a vanadium redox-flow battery (VRB). This is an already established type of energy storage cell that is very well-suited for the needs of the electrical grid as it can stay idle for very long times without losing charge, is much safer than a lithium-ion cell (though less energy-dense), is nearly immune to temperature extremes, and can be scaled up very easily, simply by increasing the size of its electrolyte tanks.

According to the researchers, the vanadium flow battery works especially well with their hybrid electrode, allowing them to boost the electric current, offering great reversibility (with 95 percent Faradaic efficiency) and allowing for high-capacity energy storage.

“We have demonstrated simultaneously reversible storage of both solar energy and electrons in the cell,” says lead author of the paper Dong Liu. “Release of the stored electrons under dark conditions continues solar energy storage, thus allowing for continuous storage around the clock.”

The team is now working on building a larger prototype, with the hope that this technology could be used to better integrate photoelectrochemical cells within the smart grid.

References:http://www.gizmag.com/