Important modern applications such as machine learning, deep learning, graph processing, databases (and many others) are memory-bound. This creates a bottleneck caused by the movement of large amounts of data between ...
详细信息
ISBN:
(纸本)9798350373769;9798350373752
Important modern applications such as machine learning, deep learning, graph processing, databases (and many others) are memory-bound. This creates a bottleneck caused by the movement of large amounts of data between the main memory and the CPU. processing-in-Memory (PiM) is currently viewed as a useful new paradigm to alleviate such bottlenecks by computing the data where it resides, i.e., in memory itself. Our goal is to analyze the potential of modern general-purpose PiM architectures to accelerate neural networks (NNs), which constantly consume high volumes of new data (i.e., low data reutilization) and are ideal for in-memory processing. We selected the UPMEM system as it is the first commercially available general-purpose PiM architecture. In this work, we implemented a multi-layer perceptron and evaluated the implementation in real use cases. We compared the PiM implementation with a sequential version running in an Intel Xeon Silver 4215 CPU. The UPMEM implementation achieved up to 260 Chi speedup for performing inference exploiting the available PiM memory.
暂无评论