LinkedIN | Twitter | GitHub | Google Scholar | CV
About me:
I am a third-year PhD student in Machine Learning at University of Washington Seattle where I am working with Prof. Jeffery A. Bilmes, part of UW ML group. I like to work on improving the efficiency of large-scale models, for both training and inference; focusing on submodular optimization for Active Learning and coreset selection.
I am also interested in robustness of large scale VLMs, and Diffusion Models. I recently interned at Amazon AWS Reckognition research team, working on diffusion models and reinforcement learning.
I completed my undergraduate degree in Electrical Engineering at IIT Delhi, where I worked with Dr Sumeet Agarwal in psycholinguistics and with Dr Prathosh AP in time series representational learning.
I'm also an avid photographer (Checkout my Unsplash) and hiker ! When not working, I can be found messing around with my camera, google earth, or hiking some beautiful mountains in North Cascades. (Most Recent Hike: Squire Pass, Boulder River Wilderness).
I can be reached at gbhatt2 [at] uw [dot] edu (or) gbhatt2 [at] cs [dot] washington [dot] edu
Recent Updates
[Mar 2024] New preprint out - Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization
[Mar 2024] New preprint out - Deep Submodular Peripteral Network
[Mar 2024] I will join NVIDIA as research intern in Summers!
[Feb 2024] "LabelBench: A Comprehensive Framework for Benchmarking Label-Efficient Learning" got accepted at DMLR Journal.
[Jan 2024] Our paper "Matryoshka Representation Learning" officially gets deployed in OpenAI new embedding models! (Click to view)
[Jan 2024] New preprint "An Experimental Design Framework for Label-Efficient Supervised Finetuning of Large Language Models" !
[Dec 2023] "Effective Backdoor Mitigation Depends on the Pre-training Objective" won the best paper award 🏆 at BUGS workshop @NeurIPS'23 (Oral)!
Publication(s)
Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization
Hritik Bansal*, Ashima Suvarna*, Gantavya Bhatt*, Nanyun Peng, Kai-Wei Chang, Aditya Grover
Under Review
An Experimental Design Framework for Label-Efficient Supervised Finetuning of Large Language Models
Gantavya Bhatt*, Yifang Chen*, Arnav Das*, Jifan Zhang*, Sang Truong, Stephen Mussmann, Yinglun Zhu, Jeff Bilmes, Simon Shaolei Du, Kevin Jamieson, Jordan P Ash, Robert D Nowak
Under Review
LabelBench: A Comprehensive Framework for Benchmarking Label-Efficient Learning
Jifan Zhang*, Yifang Chen*, Gregory Canal, Arnav Das†, Gantavya Bhatt†, Stephen Mussmann, Yinglun Zhu, Jeff Bilmes, Simon Shaolei Du, Kevin Jamieson, Robert D Nowak
In Adaptive Experimental Design and Active Learning in the Real World workshop at NeurIPS'23.
Accepted at DMLR'24
RadarHD: Demonstrating Lidar-like Point Clouds from mmWave Radar
Akarsh Prabhakara, Tao Jin, Arnav Das*, Gantavya Bhatt*, Lilly Kumari, Elahe Soltanaghei, Jeff Bilmes, Swarun Kumar, Anthony Rowe
In Annual International Conference On Mobile Computing And Networking ACM MobiCom '23
High Resolution Point Clouds from mmWave Radar
Akarsh Prabhakara, Tao Jin, Arnav Das*, Gantavya Bhatt*, Lilly Kumari, Elahe Soltanaghei, Jeff Bilmes, Swarun Kumar, Anthony Rowe
In IEEE International Conference on Robotics and Automation (ICRA'23)
Matryoshka Representations for Adaptive Deployment
Aditya Kusupati*, Gantavya Bhatt*, Aniket Rege*, Matthew Wallingford, Aditya Sinha, Vivek Ramanujan, William Howard-Snyder, Kaifeng Chen, Sham Kakade, Prateek Jain, and Ali Farhadi
In Neural Information Processing Systems (NeurIPS'22)
Systematic Generalization in Neural Networks-based Multivariate Time Series Forecasting Models
Hritik Bansal*, Gantavya Bhatt*, Pankaj Malhotra and Prathosh AP
In International joint Conference on Neural Networks (IJCNN'21)
Can RNNs trained on harder subject-verb agreement instances still perform well on easier ones?
Hritik Bansal*, Gantavya Bhatt* and Sumeet Agarwal
In Proceedings of the Society for Computation in Linguistics: Vol. 4 , Article 38.
Decay RNN
How much complexity does an RNN architecture need to learn syntax-sensitive dependencies?
Gantavya Bhatt*, Hritik Bansal*, Rishubh Singh* and Sumeet Agarwal
In Proceedings of the Society for Computation in Linguistics: Vol. 4 , Article 38.
Preprint(s)
Alignment of text-to-image models with supervised fine-tuning
Achin Jain*, Gantavya Bhatt*, Shamit Lal, Yang Zou,Yusheng Xie, Ying Wang, Ali Jahanian, Betty Mohler Tesch, R. Manmatha, Ashwin Swaminathan, Larry S. Davis, Stefano Soatto