Energy-saving designs for data-intensive computer processing

AIHub 

Researchers have demonstrated methods for both designing innovative data-centric computing hardware and co-designing hardware with machine-learning algorithms that together could improve energy efficiency by as much as two orders of magnitude. Advances in machine learning have ushered in a new era of computing -- the data-centric era -- and are forcing engineers to rethink aspects of computing architecture that have gone mostly unchallenged for 75 years. "The problem is that for large-scale deep neural networks, which are state-of-the-art for machine learning today, more than 90% of the electricity needed to run the entire system is consumed in moving data between the memory and processor," said Yingyan Lin, an assistant professor of electrical and computer engineering. Lin and collaborators proposed two complementary methods for optimizing data-centric processing, both of which were presented at the International Symposium on Computer Architecture (ISCA), a conference for new ideas and research in computer architecture. The drive for data-centric architecture is related to a problem called the von Neumann bottleneck, an inefficiency that stems from the separation of memory and processing in the computing architecture that has reigned supreme since mathematician John von Neumann developed it in 1945.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found