CLOSED Call for Papers: Special Section on Parallel and Distributed Computing Techniques for AI, ML, and DL

Share this on:
Submissions Due: 1 September 2021

Artificial intelligence (AI), machine learning (ML), and deep learning (DL) have established themselves in a multitude of domains because of their ability to process and model unstructured input data. As these fields are becoming increasingly integrated into our daily lives, there is a significant amount of interest among the community in improving AI/ML/DL through the use of parallel and distributed computing techniques (sometimes referred to as “PDC for AI/ML/DL”) as well as to apply AI/ML/DL techniques to improve traditional parallel and distributed computing systems (sometimes referred to as “AI/ML/DL for PDC”).  In this special section, we hope to bring together community research in this area into a curated selection of articles.

About TPDS special sections

TPDS has recently started a new initiative called “special sections.” Compared with regular submissions to TPDS, special sections have some differences: (1) submissions are focused on special topics of interest (similar to special issues); (2) special sections have fixed deadlines for submission and notifications; and (3) special sections have a standing committee of reviewers similar to conferences. This is the first such special section that we are planning.


The timeline for the submission and review process is as follows (all deadlines are 23:59 (11:59pm) anywhere on earth (

Round 1:

  • Submission deadline: 1 September 2021 (no extensions)
  • First-round review notification: 13 October 2021 (6 weeks for reviews)
  • Notification would be one of ACCEPT, REJECT, MAJOR REVISIONS, or MINOR REVISIONS

Round 2a (only for papers that get a minor revision in Round 1):

  • Second-round submission deadline: 27 October 2021 (2 weeks for re-submission)
  • Second-round review notification: 10 November 2021 (2 weeks for reviews)
  • Notification would be one of ACCEPT or REJECT

Round 2b (only for papers that get a major revision in Round 1):

  • Second-round submission deadline: 10 November 2021 (4 weeks for re-submission)
  • Second-round review notification: 8 December 2021 (4 weeks for reviews)
  • Notification would be one of ACCEPT, REJECT, or MINOR REVISIONS

Round 3 (only for papers that got a minor revision in Round 2b):

  • Third-round submission deadline: 22 December 2021 (2 weeks for re-submission)
  • Third-round review notification: 5 January 2022 (2 weeks for reviews)
  • Notification would be one of ACCEPT or REJECT

Topics of interest

The special section is dedicated to parallel and distributed computing (PDC) techniques for AI/ML/DL. That includes both “PDC for AI/ML/DL”- and “AI/ML/DL for PDC”-oriented articles (please see the description above). Topics of interest include, but are not limited to:

  • AI/ML/DL for PDC and PDC for AI/ML/DL
  • Data parallelism and model parallelism
  • Efficient hardware for AI, ML, and DL
  • Hardware-efficient training and inference
  • Performance modeling of AI/ML/DL applications
  • Scalable optimization methods for AI/ML/DL
  • Scalable hyper-parameter optimization
  • Scalable neural architecture search
  • Scalable IO for AI/ML/DL
  • Systems, compilers, and languages for AI/ML/DL at scale
  • Testing, debugging, and profiling AI/ML/DL applications
  • Visualization for AI/ML/DL at scale

Submission instructions

Submissions to the special section will be received as TPDS regular papers (survey and comment-style papers are not allowed). Please check submission instructions including page limit, manuscript format, and submission guidance on the TPDS Author Information page. Please note that review versions of the papers are limited to 12 pages, and overlength page charges are only for the final versions of the papers.

Submissions are *NOT* double blind. Authors can disclose their names, and they can freely cite their previous work without referring to it in a third-party fashion.

Authors can submit papers till the deadline through ScholarOne. Once you start the submission process, in Step 1 of the process, you’ll be asked to pick a “Type” for the paper. Please pick “SS for Parallel and Distributed Computing Techniques for AL, ML and DL.”

Extensions of Prior Papers

All papers need to have sufficient new content and contributions (see examples of extension material below) to warrant a separate publication. While the specific amount of acceptable new content is subjective and depends on the reviewer, we estimate that most reviewers expect new material that represents novel research contributions beyond the original publication. Acceptance of the paper is based on this new content and its contributions. Old content from previous conference papers is mainly to help reviewers understand the context. Old content should be clearly cited from the original source. Furthermore, any content used verbatim from previous publications should be appropriately quoted and cited to avoid self-plagiarism.

Authors submitting an extension of a prior publication should clearly respond to the following questions:

  1. What are the novel contributions of the submitted paper (beyond the authors’ previous publication(s))?
  2. What is the new content and in which sections does this content appear in the submission?
  3. How do the contributions (and content) build on the previous published material?

Examples of Extension Material:

Acceptable new content and contributions

  1. New conceptual extensions
  2. Experiments that provide new insights
  3. New theoretical analysis and/or proofs supporting empirical results

Allowable but insufficient content and contributions

  1. Extension to background and/or related work
  2. Elaboration on the same points in the introduction, observations, and conclusions
  3. Additional figures/plots that merely illustrate already-published content
  4. Additional experimental results without new insights.

Unacceptable content and contributions

  1. Simple union of content from multiple prior publications


  • Antonio J. Peña (Barcelona Supercomputing Center)
  • Min Si (Facebook)
  • Jidong Zhai (Tsinghua University)

Committee members

  • Junya Arai, Nippon Telegraph and Telephone Corporation, Japan
  • Jiaju Bai, Tsinghua University, China
  • Neelima Bayyapu, NITK Surathkal, India
  • Adrián Castelló, Universitat Jaume I de Castello, Spain
  • Quan Chen, Shanghai Jiaotong University, China
  • Amelie Chi Zhou, Shenzhen University, China
  • Bronis de Supinski, Lawrence Livermore National Laboratory, USA
  • Sheng Di, Argonne National Laboratory, USA
  • Nikoli Dryden, ETH Zurich, Switzerland
  • Lin Gan, Tsinghua University, China
  • Balazs Gerofi, RIKEN Center for Computational Science, Japan
  • Stephen Herbein, Laurence Livermore National Laboratory, USA
  • Zhiyi Huang, University of Otago, New Zealand
  • Sergio Iserte, Jaume I University, Spain
  • Jithin Jose, Microsoft, USA
  • Ang Li, Pacific Northwest National Laboratory, USA
  • Dong Li, University of California, Merced, USA
  • Jiajia Li, Pacific Northwest National Laboratory, USA
  • Haikun Liu, Huazhong University of Science and Technology, China
  • Weifeng Liu, China University of Petroleum-Beijing, China
  • Siyuan Ma, Google, USA
  • Naoya Maruyama, NVIDIA, USA
  • Xuehai Qian, University of Southern California, USA
  • Tong Shu, Southern Illinois University Carbondale, USA
  • Dandan Song, Beijing Institute of Technology, China
  • Wei Sun, Facebook, USA
  • Shanjiang Tang, Tianjin University, China
  • Pedro Valero Lara, Oak Ridge National Laboratory, USA
  • Hao Wang, The Ohio State University, USA
  • Haojie Wang, Tsinghua University, China
  • Zhaoguo Wang, Shanghai Jiaotong University, China
  • Zeyi Wen, The University of Western Australia, Australia
  • Rio Yokota, Tokyo Institute of Technology, Japan
  • Yang You, National University of Singapore, Singapore
  • Teng Yu, Tsinghua University, UK
  • Feng Zhang, Renmin University, China