ML Efficiency
Channel: #ml-efficiency
Co-leads:Â
Viraat - @viraat on Discord, @viraataryabumi on Twitter
Harsha - @majormelancholy on Discord, @Sree_Harsha_N on Twitter
Previous leads:
Bhavnick - @bhavnicksm on Discord, @BhavnickMinhas on Twitter
Recent Presentations
![](https://www.google.com/images/icons/product/drive-32.png)
June 21, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
May 24, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
May 10, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
April 12, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
March 1, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
February 19, 2024
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
Join us to explore the theme of Transformer Inference Optimization.
After last session's paper on LLM.int8(), this week we will be going through "SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression" by Dettmers et. al.
The wonderful @Srishti Gureja will be presenting again!
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
![](https://www.google.com/images/icons/product/drive-32.png)
On the Efficacy of Knowledge Distillation" by Hariharan et. al (2019)
Slides
![](https://www.google.com/images/icons/product/drive-32.png)
Knowledge Distillation and the paper: Model Compression by Caruana et. al.
Materials from all past sessions