LinkedIN | Twitter | Email | Google Scholar | CV
I am a fourth-year PhD student at University of Washington Seattle where I am working with Prof. Jeffery A. Bilmes. I like to work on improving the data efficiency of large-scale models, for both training and inference; with the help of submodular optimization. In Summer'25, I am interning at NVIDIA - Applied Deep Learning Research working with Mohammad Shoeybi and Mostafa Patwary. In summer'24 I did an internship at NVIDIA - Applied Deep Learning Research Team with Rafael Valle and Kevin Shih.
Before joining PhD program, I completed my undergraduate degree at IIT Delhi, where I worked with Dr Sumeet Agarwal and Dr Prathosh AP. I'm also an avid photographer (Checkout my Unsplash) and hiker! When not working, I can be found messing around with my camera, google earth, or hiking in North Cascades National Park.
[June 2025] Joined NVIDIA as Research Scientist Intern.
[June 2025] New preprint on Submodular Distribution Matching with Sparse Features for Multimodal Data Filtering
[May 2025] Aligning Large Language Models via Joint Preference Optimization is accepted in ACL'25!
[Feb 2025] COBRA is accepted in CVPR'25. Hi Nashville, TN!
[Oct 2024] New preprint How Many Van Goghs Does It Take to Van Gogh? Finding the Imitation Threshold.
[Sept 2024] Deep Submodular Peripteral Networks to appear at NeurIPS'24 as spotlight!
[June 2024] 2 Papers (COBRA and DOVE) will appear in DMLR workshop for ICML'24 as poster and oral talk, respectively.
[June 2024] DOVE will appear in Models of Human Feedback for AI Alignment workshop for ICML'24 as a poster.
[May 2024] New preprint COBRA: COmBinatorial Retrieval Augmentation for Few-Shot Learning.
[May 2024] "An Experimental Design Framework for Label-Efficient Supervised Finetuning of Large Language Models" got accepted at ACL Findings.
Submodular Distribution Matching w/ Sparse Features for Multimodal Data Filtering
Arnav Das*, Gantavya Bhatt*, Yiping Wang, Viswa Virinchi, Sahil Verma, Simon Shaolei Du, Jeff Bilmes
Under Review
How Many Van Goghs Does It Take to Van Gogh? Finding the Imitation Threshold
Sahil Verma, Royi Rassin, Arnav Das*, Gantavya Bhatt*, Preethi Seshadri*, Chirag Shah, Jeff Bilmes, Hannaneh Hajishirzi, Yanai Elazar
[Invited Talk] RegML workshop at NeurIPS 2024
[Poster] Accepted at ATTRIB, RegML, and SafeGenAI workshops at NeurIPS 2024 and NLLP Workshop 2024
Under Review
COBRA: COmBinatorial Retrieval Augmentation for Few-Shot Learning
Arnav Das*, Gantavya Bhatt*, Lilly Kumari, Sahil Verma, Jeff Bilmes
[Poster] In Conference on Computer Vision and Pattern Recognition (CVPR'25)
[Poster] In DMLR workshop at ICML'24
Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization
Hritik Bansal*, Ashima Suvarna*, Gantavya Bhatt*, Nanyun Peng, Kai-Wei Chang, Aditya Grover
[Poster] Findings of ACL'25.
[Oral] In DMLR workshop at ICML'24
[Poster] In MHFAI Alignment workshop at ICML'24
Deep Submodular Peripteral Networks
Gantavya Bhatt*, Arnav Das*, Jeff Bilmes
[Spotlight] In Neural Information Processing Systems (NeurIPS'24)
An Experimental Design Framework for Label-Efficient Supervised Finetuning of Large Language Models
Gantavya Bhatt*, Yifang Chen*, Arnav Das*, Jifan Zhang*, Sang Truong, Stephen Mussmann, Yinglun Zhu, Jeff Bilmes, Simon Shaolei Du, Kevin Jamieson, Jordan P Ash, Robert D Nowak
[Poster] Accepted at ACL'24 (Findings)
Effective Backdoor Mitigation Depends on the Pre-training Objective
Sahil Verma, Gantavya Bhatt, Avi Schwarzschild, Soumye Singhal, Arnav Das, Chirag Shah, John P Dickerson, Jeff Bilmes
In Transactions of Machine Learning Research (Jan'25 edition, TMLR)
[Best Paper Award 🏆] In BUGS workshop at NeurIPS'23
Under review
LabelBench: A Comprehensive Framework for Benchmarking Label-Efficient Learning
Jifan Zhang*, Yifang Chen*, Gregory Canal, Arnav Das†, Gantavya Bhatt†, Stephen Mussmann, Yinglun Zhu, Jeff Bilmes, Simon Shaolei Du, Kevin Jamieson, Robert D Nowak
In Adaptive Experimental Design and Active Learning in the Real World workshop at NeurIPS'23.
Accepted at DMLR'24
RadarHD: Demonstrating Lidar-like Point Clouds from mmWave Radar
Akarsh Prabhakara, Tao Jin, Arnav Das*, Gantavya Bhatt*, Lilly Kumari, Elahe Soltanaghei, Jeff Bilmes, Swarun Kumar, Anthony Rowe
In Annual International Conference On Mobile Computing And Networking ACM MobiCom '23
High Resolution Point Clouds from mmWave Radar
Akarsh Prabhakara, Tao Jin, Arnav Das*, Gantavya Bhatt*, Lilly Kumari, Elahe Soltanaghei, Jeff Bilmes, Swarun Kumar, Anthony Rowe
In IEEE International Conference on Robotics and Automation (ICRA'23)
Matryoshka Representation Learning
Aditya Kusupati*, Gantavya Bhatt*, Aniket Rege*, Matthew Wallingford, Aditya Sinha, Vivek Ramanujan, William Howard-Snyder, Kaifeng Chen, Sham Kakade, Prateek Jain, and Ali Farhadi
[Poster] In Neural Information Processing Systems (NeurIPS'22)
Systematic Generalization in Neural Networks-based Multivariate Time Series Forecasting Models
Hritik Bansal*, Gantavya Bhatt*, Pankaj Malhotra and Prathosh AP
In International joint Conference on Neural Networks (IJCNN'21)
Can RNNs trained on harder subject-verb agreement instances still perform well on easier ones?
Hritik Bansal*, Gantavya Bhatt* and Sumeet Agarwal
In Proceedings of the Society for Computation in Linguistics: Vol. 4 , Article 38.
Decay RNN
How much complexity does an RNN architecture need to learn syntax-sensitive dependencies?
Gantavya Bhatt*, Hritik Bansal*, Rishubh Singh* and Sumeet Agarwal
In Proceedings of the Society for Computation in Linguistics: Vol. 4 , Article 38.