Samsung announces high bandwidth memory, processing-in-memory architecture

Samsung announces high bandwidth memory, processing-in-memory architecture
Credit: Samsung

Samsung Electronics has announced on its Newsroom webpage the development of a new kind of memory chip architecture called high-bandwidth memory, processing-in-memory—HBM-PIM. The architecture adds artificial intelligence processing to high-bandwidth memory chips. The new chips will be marketed as a way to speed up data centers, boost speed in high performance computers and to further enable AI applications.

Computer engineers have long been working to remove the bottlenecks that arise in computers due to the need to shuffle data back and forth between a computer’s CPU and its memory chips. Most efforts to do so have involved adding small amounts of fast memory caches to the CPUs—unfortunately, doing so increases energy consumption, leading to more heat production. In this new effort, the team at Samsung has taken the opposite approach—giving memory chips the ability to take on some of the processing. With the new HBM-PIM, Samsung has placed what it describes as “a DRAM optimized AI engine inside of a memory bank.” This reduces the processing load on the CPU by offloading some of its work to remote databanks. Not only is the workload reduced, but the speed of processing is increased due to a reduction in data movement.

Specs for the HBM-PIM include a PCU running at 300MHz controlled by the host CPU using conventional memory commands. With this approach, the PCU can be instructed to carry out FP16 calculations directly inside of the DRAM unit. Notably, the HBM-PIM can operate as normal RAM when a system is running applications that have not been written for it.

Samsung notes that when they tested the new technology with their existing HBM2 Aquabolt systems, system performance doubled and energy consumption was reduced by 70%. They also noted that installing HBM-PIMs in existing systems would not require any other changes to in-place hardware or software. Their HBM-PIM technology is currently being tested with previously developed AI accelerators created by AI solution partners—they expect to see the results in the first half of this year.

Representatives of the company will be presenting a paper they have written describing the new technology at this year’s International Solid-State Circuits Virtual Conference.

More information: news.samsung.com/global/samsun … -ai-processing-power
Citation: Samsung announces high bandwidth memory, processing-in-memory architecture (2021, February 18) retrieved 18 February 2021 from https://techxplore.com/news/2021-02-samsung-high-bandwidth-memory-processing-in-memory.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Source: TechExplore