IEEE Transactions on Parallel and Distributed Systems

IEEE Transactions on Parallel and Distributed Systems (TPDS) is a scholarly archival journal published monthly. Parallelism and distributed computing are foundational research and technology to rapidly advance computer systems and their applications. Read the full scope of TPDS.

Expand your horizons with Colloquium, a monthly survey of abstracts from all CS transactions! Replaces OnlinePlus in January 2017.

From the June 2018 Issue

Towards Memory-Efficient Allocation of CNNs on Processing-in-Memory Architecture

By Yi Wang, Weixuan Chen, Jing Yang, and Tao Li

Free Featured Article Convolutional neural networks (CNNs) have been successfully applied in artificial intelligent systems to perform sensory processing, sequence learning, and image processing. In contrast to conventional computing-centric applications, CNNs are known to be both computationally and memory intensive. The computational and memory resources of CNN applications are mixed together in the network weights. This incurs a significant amount of data movement, especially for high-dimensional convolutions. The emerging Processing-in-Memory (PIM) alleviates this memory bottleneck by integrating both processing elements and memory into a 3D-stacked architecture. Although this architecture can offer fast near-data processing to reduce data movement, memory is still a limiting factor of the entire system. We observe that an unsolved key challenge is how to efficiently allocate convolutions to 3D-stacked PIM to combine the advantages of both neural and computational processing. This paper presents MemoNet, a memory-efficient data allocation strategy for convolutional neural networks on 3D PIM architecture. MemoNet offers fine-grained parallelism that can fully exploit the computational power of PIM architecture. The objective is to capture the characteristics of neural network applications and perfectly match the underlining hardware resources provided by PIM, resulting in a hardware-independent design to transparently allocate data. We formulate the target problem as a dynamic programming model and present an optimal solution. To demonstrate the viability of the proposed MemoNet, we conduct a set of experiments using a variety of realistic convolutional neural network applications. The extensive evaluations show that, MemoNet can significantly improve the performance and the cache utilization compared to representative schemes.

download PDF View the PDF of this article      csdl View this issue in the digital library

Editorials and Announcements


  • We are pleased to announce that Manish Parashar, a Distinguished Professor of Computer Science at Rutgers, The State University of New Jersey University, has been selected as the new Editor-in-Chief of the IEEE Transactions on Parallel and Distributed Systems starting in 2018.
  • We are pleased to announce that Xian-He Sun, a Distinguished Professor of Computer Science at The Illinois Institute of Technology, has been selected as the new Associate Editor-in-Chief of the IEEE Transactions on Parallel and Distributed Systems starting in 2018.
  • TPDS now offers authors access to Code Ocean. Code Ocean is a cloud-based executable research platform that allows authors to share their algorithms in an effort to make the world’s scientific code more open and reproducible. Learn more or sign up for free.
  • According to Clarivate Analytics' 2016 Journal Citation Report, TPDS has an impact factor of 4.181.


Guest Editorials

Reviewers List

Annual Index

Access recently published TPDS articles

RSS Subscribe to the RSS feed of recently published TPDS content

mail icon Sign up for e-mail notifications through IEEE Xplore Content Alerts

preprints icon View TPDS preprints in the Computer Society Digital Library

TPDS is indexed in ISI