Architecture and System Support for Transformer Models

Call for Papers

Transformer models have become the foundation of a new wave of machine learning models. The application of such models spans from natural language understanding into image processing, protein folding, and many more. The main objective of this workshop is to bring the attention of our community to the upcoming architecture and system challenges for these foundational models and drive the innovation for supporting efficient execution of these ever-scaling models. To achieve this, the format of the workshop will consists of a combination of keynote speakers, short talks, followed by a panel discussion. Subject areas of the workshop included (but not limited to):

Submission Instructions

Organizing Committee

Contact us at archsystm@gmail.com