Subscribe to our free daily newsletters
. 24/7 Space News .

Subscribe to our free daily newsletters

Selective memory makes data caches 50 percent more efficient
by Staff Writers
Boston MA (SPX) Oct 26, 2017

illustration only

In a traditional computer, a microprocessor is mounted on a "package," a small circuit board with a grid of electrical leads on its bottom. The package snaps into the computer's motherboard, and data travels between the processor and the computer's main memory bank through the leads.

As processors' transistor counts have gone up, the relatively slow connection between the processor and main memory has become the chief impediment to improving computers' performance. So, in the past few years, chip manufacturers have started putting dynamic random-access memory - or DRAM, the type of memory traditionally used for main memory - right on the chip package.

The natural way to use that memory is as a high-capacity cache, a fast, local store of frequently used data. But DRAM is fundamentally different from the type of memory typically used for on-chip caches, and existing cache-management schemes don't use it efficiently.

At the recent IEEE/ACM International Symposium on Microarchitecture, researchers from MIT, Intel, and ETH Zurich presented a new cache-management scheme that improves the data rate of in-package DRAM caches by 33 to 50 percent.

"The bandwidth in this in-package DRAM can be five times higher than off-package DRAM," says Xiangyao Yu, a postdoc in MIT's Computer Science and Artificial Intelligence Laboratory and first author on the new paper. "But it turns out that previous schemes spend too much traffic accessing metadata or moving data between in- and off-package DRAM, not really accessing data, and they waste a lot of bandwidth. The performance is not the best you can get from this new technology."

Cache hash
By "metadata," Yu means data that describe where data in the cache comes from. In a modern computer chip, when a processor needs a particular chunk of data, it will check its local caches to see if the data is already there. Data in the caches is "tagged" with the addresses in main memory from which it is drawn; the tags are the metadata.

A typical on-chip cache might have room enough for 64,000 data items with 64,000 tags. Obviously, a processor doesn't want to search all 64,000 entries for the one that it's interested in. So cache systems usually organize data using something called a "hash table." When a processor seeks data with a particular tag, it first feeds the tag to a hash function, which processes it in a prescribed way to produce a new number. That number designates a slot in a table of data, which is where the processor looks for the item it's interested in.

The point of a hash function is that very similar inputs produce very different outputs. That way, if a processor is relying heavily on data from a narrow range of addresses - if, for instance, it's performing a complicated operation on one section of a large image - that data is spaced out across the cache so as not to cause a logjam at a single location.

Hash functions can, however, produce the same output for different inputs, which is all the more likely if they have to handle a wide range of possible inputs, as caching schemes do. So a cache's hash table will often store two or three data items under the same hash index. Searching two or three items for a given tag, however, is much better than searching 64,000.

Dumb memory
Here's where the difference between DRAM and SRAM, the technology used in standard caches, comes in. For every bit of data it stores, SRAM uses six transistors. DRAM uses one, which means that it's much more space-efficient. But SRAM has some built-in processing capacity, and DRAM doesn't. If a processor wants to search an SRAM cache for a data item, it sends the tag to the cache. The SRAM circuit itself compares the tag to those of the items stored at the corresponding hash location and, if it gets a match, returns the associated data.

DRAM, by contrast, can't do anything but transmit requested data. So the processor would request the first tag stored at a given hash location and, if it's a match, send a second request for the associated data. If it's not a match, it will request the second stored tag, and if that's not a match, the third, and so on, until it either finds the data it wants or gives up and goes to main memory.

In-package DRAM may have a lot of bandwidth, but this process squanders it. Yu and his colleagues - Srinivas Devadas, the Edwin Sibley Webster Professor of Electrical Engineering and Computer Science at MIT; Christopher Hughes and Nadathur Satish of Intel; and Onur Mutlu of ETH Zurich - avoid all that metadata transfer with a slight modification of a memory management system found in most modern chips.

Any program running on a computer chip has to manage its own memory use, and it's generally handy to let the program act as if it has its own dedicated memory store. But in fact, multiple programs are usually running on the same chip at once, and they're all sending data to main memory at the same time. So each core, or processing unit, in a chip usually has a table that maps the virtual addresses used by individual programs to the actual addresses of data stored in main memory.

Look here
Yu and his colleagues' new system, dubbed Banshee, adds three bits of data to each entry in the table. One bit indicates whether the data at that virtual address can be found in the DRAM cache, and the other two indicate its location relative to any other data items with the same hash index.

"In the entry, you need to have the physical address, you need to have the virtual address, and you have some other data," Yu says. "That's already almost 100 bits. So three extra bits is a pretty small overhead."

There's one problem with this approach that Banshee also has to address. If one of a chip's cores pulls a data item into the DRAM cache, the other cores won't know about it. Sending messages to all of a chip's cores every time any one of them updates the cache consumes a good deal of time and bandwidth. So Banshee introduces another small circuit, called a tag buffer, where any given core can record the new location of a data item it caches.

Any request sent to either the DRAM cache or main memory by any core first passes through the tag buffer, which checks to see whether the requested tag is one whose location has been remapped. Only when the buffer fills up does Banshee notify all the chips' cores that they need to update their virtual-memory tables. Then it clears the buffer and starts over.

The buffer is small, only 5 kilobytes, so its addition would not use up too much valuable on-chip real estate. And the researchers' simulations show that the time required for one additional address lookup per memory access is trivial compared to the bandwidth savings Banshee affords.

Fast-moving magnetic particles could enable new form of data storage
Boston MA (SPX) Oct 05, 2017
New research has shown that an exotic kind of magnetic behavior discovered just a few years ago holds great promise as a way of storing data - one that could overcome fundamental limits that might otherwise be signaling the end of "Moore's Law," which describes the ongoing improvements in computation and data storage over recent decades. Rather than reading and writing data one bit at a ti ... read more

Related Links
Massachusetts Institute of Technology
Space Technology News - Applications and Research

Thanks for being here;
We need your help. The SpaceDaily news network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.

SpaceDaily Contributor
$5 Billed Once

credit card or paypal
SpaceDaily Monthly Supporter
$5 Billed Monthly

paypal only

Comment using your Disqus, Facebook, Google or Twitter login.

Share this article via these popular social media networks DiggDigg RedditReddit GoogleGoogle

Plants and psychological well-being in space

Spacewalkers fix robotic arm in time to grab next cargo ship

NASA develops and tests new housing for in-orbit science payloads

Russia's space agency says glitch in manned Soyuz landing

NASA awards launch contracts for Landsat 9 and Sentinel-6A

ESA role in Europe's first all-electric telecom satellite

Lockheed Martin Launches Second Cycle of 'Girls' Rocketry Challenge' in Japan

First Four Space Launch System Flight Engines Ready To Rumble

Solar eruptions could electrify Martian moons

MAVEN finds Mars has a twisted tail

Mine craft for Mars

Opportunity spends the week imaging Perseverance Valley

China launches three satellites

Mars probe to carry 13 types of payload on 2020 mission

UN official commends China's role in space cooperation

China's cargo spacecraft separates from Tiangong-2 space lab

Myanmar to launch own satellite system-2 in 2019: vice president

Eutelsat's Airbus-built full electric EUTELSAT 172B satellite reaches geostationary orbit

Turkey, Russia to Enhance Cooperation in the Field of Space Technologies

SpaceX launches 10 satellites for Iridium mobile network

The drop that's good to the very end

Study shows how rough microparticles can cause big problems

Xenesis Licenses Cutting Edge IP from NASA/JPL

These headsets are made for walking over Mars

New NASA study improves search for habitable worlds

From Comets Come Planets

A star that devoured its own planets

Astronomers find potential solution into how planets form

Haumea, the most peculiar of Pluto companions, has a ring around it

Ring around a dwarf planet detected

Helicopter test for Jupiter icy moons radar

Solving the Mystery of Pluto's Giant Blades of Ice

Memory Foam Mattress Review
Newsletters :: SpaceDaily :: SpaceWar :: TerraDaily :: Energy Daily
XML Feeds :: Space News :: Earth News :: War News :: Solar Energy News

The content herein, unless otherwise known to be public domain, are Copyright 1995-2017 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement