The 1st Workshop on Reimagining Distributed Computing for LLMs (DistLLM) aims to tackle the pressing challenges in the scalable, efficient, and reliable distributed training, fine-tuning, and inference of large language models (LLMs) and deep neural networks (DNNs). This workshop brings together leading experts in distributed systems, networking, and AI to collaboratively explore innovative solutions across multiple critical domains. These include advanced multi-dimensional parallelism techniques encompassing data, tensor, pipeline, and expert parallelism; communication and synchronization protocols optimized for efficient collective communication and congestion control; and intelligent distributed scheduling strategies that address multi-tenant resource allocation and adaptive scheduling under communication bottlenecks.
Further, the workshop emphasizes network-aware system design to ensure load-balanced traffic and topology-aware flow control, alongside developing resilience and fault tolerance methods such as checkpointing and anomaly detection for long-running training jobs. Recognizing the growing complexity of LLM training infrastructure, the event also addresses emerging challenges of multi-cluster training spanning data centers and edge clusters, including inter-domain and wide area network scheduling. Additionally, specialized topics on LLM deployment tailored for edge computing environments—such as fine-tuning tiny LLMs and optimizing for heterogeneous compute devices—are featured.
Through a comprehensive program of refereed papers, keynote talks, expert panels, lightning talks, and demonstrations, the DistLLM workshop seeks to advance the state-of-the-art in distributed systems infrastructure for AI while leveraging LLM capabilities to innovate distributed computing itself. By fostering interdisciplinary collaboration, this workshop intends to drive transformative research that meets the extraordinary demands of modern AI workloads and bridges key gaps in scalable, efficient, and resilient distributed model training and deployment.
Parallelism-aware scheduling and job orchestration
Scalable parallel algorithms for data, model, pipeline, and expert parallelism
Collective protocols and congestion mitigation
Fault tolerance, checkpointing, and recovery for long-running training
Multi-tenant cluster scheduling and resource sharing
Geo-distributed and edge-coordinated model training
System design for MoE, long-sequence models, and sequence parallelism
Tools and frameworks for visualizing, debugging, or optimizing training systems
Security, Privacy, and Isolation-aware distributed training
Protocols for consistent and efficient tracking of model checkpoints
Algorithms for fair and concurrent training of multiple models
LLMs for edge compute nodes
Important Dates: (All Dates are Anywhere on Earth)
Submission deadline: September 15th, 2025
Notification to authors: October 25th, 2025
Camera-ready paper due: November 15th, 2025
Workshop date: January 6-9th, 2026 (Nara, Japan) [Exact date to be announced]
All papers must be original and not simultaneously submitted to another journal or conference. All papers will be peer-reviewed using a double-blind peer-review process by at least three members of the program committee. Submissions should be a complete manuscript. DistLLM accepts Full papers: 6 pages in ACM Conference format (including title, abstract, figures, and references).
Papers should be formatted in double-column, single-spaced layout using a 10-point font size on standard 8.5 x 11-inch (US letter) pages.
Submissions are anonymous. The conference will employ a lightweight double-blind reviewing process. Manuscripts should not include author names and affiliations.
Submissions should not reveal the identity of the authors in any way. Authors should ensure that any references to their own related work are in the third person (e.g., not “We build on our previous work …” but rather “We build on the work of …”).
Authors are required to use the official ACM conference templates for manuscript preparation, available in both MS Word and LaTeX formats.
ACM templates ensure compliance with ACM’s publication standards and can be found here:
https://www.acm.org/publications/proceedings-template
Papers are to be submitted electronically in PDF format. Submitted papers should not have appeared in or be under consideration for a different workshop, conference or journal. All accepted papers need to be presented at the workshop by one of the authors.
All accepted papers (subject to post-review revisions) will be published in the ICDCN 2026 companion proceedings.
Submission Link: https://easychair.org/conferences?conf=icdcn2026
Select WS2: 1st Workshop on Reimagining Distributed Computing for LLMs for submissions.
Here are the guidelines for Artificial Intelligence (AI)-Generated Text and related policies for the workshop papers, based on the provided ACM policies and the instructions given:
Authorship and Use of Generative AI:
Ensure that the ACM Policy on Authorship is followed strictly for all accepted papers. This means all authors must be identifiable human beings who made substantial intellectual contributions and take responsibility for the work.
Generative AI tools and technologies such as ChatGPT may be used to assist in creating sections of the work (text, code, data, citations, etc.), but these must be fully disclosed in the acknowledgements section of the paper.
Basic word processing assistance (spell check, grammar correction) does not require disclosure.
Authorship cannot be added or removed after paper acceptance.
Open Publication Model and Article Processing Charges (APC):
ICDCN 2026 workshop proceedings will be published as a companion volume along with the main conference proceedings. However, please note that ACM has moved to a new open-access publishing model for all conference proceedings to be published via ACM ICPS. The authors have to pay an Article Processing Charge (APC) to ACM (which is beyond the regular conference registration fee) if the corresponding author’s organization is not a member of the ACM Open program. Please check the details through this link: https://www.acm.org/publications/icps/author-guidance. Several institutes worldwide are already members of the ACM Open program. The authors can use the following link to check if their organization is a member under the ACM Open program: https://libraries.acm.org/acmopen/open-participants. For any clarifications, contact icps-info@acm.org.
Here are the specific ACM links related to the policies and guidelines mentioned for the workshop papers:
ACM Policy on Authorship:
https://www.acm.org/publications/policies/new-acm-policy-on-authorship
ACM Policy on the use of Generative AI in papers (FAQ):
https://www.acm.org/publications/policies/frequently-asked-questions
ACM Open Publication Model and FAQs for International Conference Proceedings Series (ICPS):
https://www.acm.org/publications/icps/faq
List of Institutions Participating in the ACM Open Program:
https://libraries.acm.org/acmopen/open-participants
Guidance on Article Processing Charges (APC) and publishing in the ACM Digital Library:
https://www.acm.org/publications/icps/author-guidance
ICDCN 2026 Main Conference Paper Submission Guidelines (including double-blind review process):
https://sites.google.com/view/icdcn2026/submissions/call-for-paper?authuser=0
These links will help authors and organizers ensure compliance with ACM's policies on authorship, generative AI usage, open access publishing, and paper submission guidelines.