Welcome to Rishabh Iyer's webpage

Assistant Professor, University of Texas Dallas

I am currently an Assistant Professor at the University of Texas, Dallas, where I lead the CARAML Lab. I'm also a Visiting Assistant Professor at the Indian Institute of Technology, Bombay. Before this, I was a Senior Research Scientist at Microsoft between 2016 till 2019. Below are some of the areas my group is currently working on:

  • Compute-Efficient Learning via data subset selection, and coresets: obtaining 5x - 10x speedups/energy efficiency with small data subsets with negligible loss in accuracy (generalization performance) applied to varied applications like supervised, semi-supervised, hyper-parameter tuning, NAS and domains like image classification, NLP, speech recognition.

  • Data-Efficient Learning and Active Learning: learning with fewer labels, reducing labeling costs by 2x - 5x (see this blog page on some of our work in active learning), particularly in realistic scenarios like rare classes/slices, imbalance, OOD instances, redundancy, etc.

  • Robust Learning in the presence of outliers, noise etc.

  • Data Summarization: Video/Image/Text (summarize massive datasets with scalable discrete optimization)

  • Model Compression/Pruning, Feature Selection, Cost-sensitive Feature Selection (reduce model size for deployment in resource-constrained environments)

  • Learning with Rules, Labeling Functions, and Data Programming

  • Discrete Optimization (specifically submodular optimization)

  • Combinatorial (Submodular) Information Measures

Our research is currently supported by grants from NSF, Adobe Data Science Award, a Google Gift, Amazon Research Award, and the UT Dallas Seed grant. Thank you! Our research is motivated by real-world problems in machine learning, computer vision, text, and NLP! For more on my research, please see my research page, my publications, or my lab webpage.

I completed my Ph.D. in 2015 from the University of Washington, Seattle where I worked with Jeff Bilmes. I am excited about making machines assist humans in processing massive amounts of data, particularly in understanding videos and images. I am interested in building intelligent systems which organize, analyze and summarize massive amounts of data, and also automatically learn from this.

I received the best paper awards at Neural Information Processing Systems (NeurIPS/NIPS) in 2013, the International Conference of Machine Learning (ICML) in 2013, and an Honorable Mention at CODS-COMAD in 2021. I also received several research awards including an NSF Medium Grant, an Adobe Data Science Research Award, Microsoft Research Ph.D. Fellowship, Facebook Ph.D. Fellowship, and the Yang Award for Outstanding Graduate Student from the University of Washington.

For more information, please see my Google Scholar Profile, LinkedIn Profile, DBLP, or my GitHub page. I also maintain a YouTube channel where I add videos of my lectures and research talks.

Twitter: @rishiyer

Awards and Recognition

  • Amazon Research Award on Fairness for AI, July 2022 (Main PI: Preethi Jyothi)

  • Adobe Data Science Research Award, February 2022

  • NSF Medium Grant, September 2021.

  • Honorable Mention for our paper at CODS-COMAD 2021

  • Outstanding Reviewer Award for NeurIPS 2020 and 2021!

  • Finalist in the LDV Computer Vision Conference, New York in 2017

  • Yang Outstanding Graduate Student Award, University of Washington, Seattle

  • Microsoft Research Fellowship Award, 2014

  • Facebook Fellowship Award. 2014 (Declined in favor of Microsoft)

  • Best Paper Award at the International Conference of Machine Learning, 2013

  • Best Paper Award at the Neural Information Processing Systems Conference, 2013

Work Experience and Education

  • Spring 2020 to Present, Assistant Professor at the CS Department, UT Dallas

  • August 2020 to Present, Visiting Assistant Professor at CSE Department, IIT Bombay

  • March 2016 - December 2019, Senior Research Scientist, Microsoft

  • March 2015 - March 2016, Post-Doctoral Researcher, University of Washington

  • September 2011 - March 2015, M.S and Ph.D., University of Washington, Seattle

  • August 2011 - May 2011, B.Tech, IIT-Bombay


Recent news

  • Two papers: ORIENT on data subset selection for distribution shift, and AUTOMATA on gradient-based subset selection for compute-efficient hyper-parameter tuning are accepted at NeurIPS 2022!

  • Our work on Robust Semi-supervised Learning was accepted at ICDM 2022! Congrats Krishnateja and Xujiang!

  • Excited to Recieve an Amazon Research Award in Alexa Fairness for AI track for our proposal on Fair Speech Recognition using Targeted Subset Selection and Active Semi-supervised Learning (joint with Preethi Jyothi and Ganesh Ramakrishnan from IIT Bombay).

  • Our work TALISMAN, which is on Targeted Active Learning for Object Detection in Autonomous Driving with Rare Scenarios (e.g, Mining instances of "Motorcycles at night" or "Pedestrians on a Highway in Foggy Scenarios") accepted at ECCV 2022!

  • Our work on Semi-supervised Meta-Learning (PLATINUM) was accepted to ICML 2022!

  • Excited to receive the Adobe Data Science Research Award for February 2022! Thanks, Adobe for the support!

  • Our work on gradient coresets for continuous learning was accepted at CVPR 2022 and our work on robust learning of labeling functions was accepted at Findings of ACL 2022! Congrats Krishnateja and all our coauthors!

  • Invited Speaker at the "Bi-level Optimization in Machine Learning" and "Submodular Optimization" sessions at INFORMS Optimization Society (IOS) 2022!

  • Two papers from CARAML lab are accepted at AAAI 2022! Congrats Krishnateja and Suraj!

  • Our work on Submodular Information Measures was accepted to Transactions of Information Theory Journal.

  • Outstanding Reviewer Award from NeurIPS 2021! (received the same award in 2020 as well).

  • Excited to be giving a tutorial at AAAI 2022 on Subset Selection in Machine Learning: Theory, Applications, and Hands-on. Stay tuned for more updates!

  • Three papers from CARAML lab are accepted at NeurIPS 2021! Congrats Krishna, Ping, Nathan, and Suraj!

  • Received an NSF Collaborative Medium Award on Submodular Information Functions with Applications to Machine Learning. Thanks, NSF!

  • Received gift funding from Adobe for Targeted Subset Selection! Thanks, Adobe!

  • Received gift funding from Google on Continuous Learning! Thanks, Google!

  • Together with Abir De, Ganesh Ramakrishnan, and Jeff Bilmes, I am co-organizing a workshop on Subset Selection in Machine Learning: From Theory to Applications at ICML 2021 on July 24th 2021! Workshop page: https://icml.cc/virtual/2021/workshop/8351

  • Will serve as an Area chair for AAAI 2022

  • Excited to release submodlib (Github: https://github.com/decile-team/submodlib), a submodular optimization toolkit. Credits to Vishal Kaushal for leading this effort.

  • Excited to release CORDS (Github: https://github.com/decile-team/cords), a PyTorch-based open-source efficient deep model training and autoML library! Credits to my student Krishnateja Killamsetty for leading this.

  • Excited to release DISTIL (Github: https://github.com/decile-team/distil), a PyTorch-based open-source active learning toolkit for deep learning! Credits to my students Nathan Beck and Durga Sivasubramanian for leading this.

  • Two papers (GRAD-MATCH and SELCON) accepted to ICML 2021!

  • Two papers on rule augmented learning accepted at Findings of ACL 2021 (one short and one long).

  • Invited Speaker at the London Symposium on Information Theory (LSIT) 2021 (Youtube Link to the Talk)

  • Happy to announce that we have released VISIOCITY, a dataset comprising of long videos for video summarization, and more broadly video understanding!

  • Our work on "A Clustering based Selection Framework for Cost Aware and Test-time Feature Elicitation" received Best Paper Honorable Mention at CODS-COMAD 2021! Congrats Srijita and Sriraam!

  • I will be presenting a tutorial on Combinatorial Approaches for Data, Topic and Feature Selection and Summarization at IJCAI 2020 with Ganesh Ramakrishnan (presented a similar one at ECAI 2020 earlier this year).

  • Our paper on Combinatorial Information Measures accepted for publication at ALT 2021, and our paper on Online Video Summarization accepted at SDM 2021!

  • Our paper on Data Subset Selection (GLISTER) accepted to AAAI 2021!

  • Invited Talk in the Special Session Deep Learning and Information Theory at SPCOM 2020 (Virtual)

  • Senior Program Committee for AAAI 2021

  • Selected among the 10% of the Reviewers for NeurIPS 2020

  • Our paper on Concave Aspects of Submodular Functions accepted at ISIT 2020.

  • Invited Speaker and Participant at the Workshop on Optimization in Machine Learning at IST Austria, May 2020

  • Invited Speaker at the Information Theory and Applications (ITA) workshop at San Diego, CA in February 2020

  • Our paper on Robust Submodular Minimization accepted at ECAI 2020!

  • I'm teaching (and designing) a new course at UT Dallas in Spring 2020 on Optimization in Machine Learning (Course Website). This course will cover the basics of both continuous and discrete optimization in ML. This course will be a mix of theory and practical (implementational) aspects of continuous and discrete optimization.

  • I've joined the CS Department of University of Texas, Dallas in Spring 2020 as an Assistant Professor.

  • September 2019: Visited Tata Institute of Fundamental Research for an Invited Talk

  • February 2019: Visited University of Texas at Dallas and University of Pittsburgh in February 2018 and gave a talk on Scalable and Practical Discrete Optimization for Big Data (see this link).

  • December 2018: Two papers accepted into AISTATS 2019!

  • October 2018: Tutorial Speaker at the 7th IEEE Winter Conference on Applications of Computer Vision (WACV) 2019 (see tutorial website. Slides are on the website)

  • October 2018: Three papers accepted to WACV 2019!

  • October 2018: Video Analytics software developed with collaborators at IIT Bombay available now at this link.

  • October 2018: Invited Talk at Allen Institute of AI and Google Seattle, October 2018 (Video Link)

  • July 2018: Released Open Source software Jensen with my collaborators John Halloran and Kai Wei

  • May 2017: Presented our work on Online Learning for Click Prediction at the Microsoft Machine Learning, AI and Data Science Conference

  • May 2017: Finalist at the LDV Vision Conference, New York

  • March 2017: Invited Speaker at AMS Sectional Meeting, Special Session on Geometry and Optimization in Computer Vision, Pullman, WA

  • March 2017: Our work on Limited Vocabulary Speech Data Subset Selection selected to Appear in Computer Speech & Language, 2017. Corpus Definitions and Baselines for SVitchboard-II and FiSVer-I datasets can be found at this link.

  • April 2016: Work on Minimizing Ratio of Submodular Function accepted at ICML 2016

  • Feb 2016: Finished my PostDoc. Will be Joining Microsoft, starting March 2016.

  • Two Papers accepted in NIPS 2015, Two Papers in AISTATS 2015, One Paper in ACL and INTERSPEECH 2015 and one paper in ICML 2015

  • Invited Speaker at the International Symposium on Mathematical Programming (ISMP), Pittsburg - July, 2015 (Session on Submodular Optimization, Link)

  • Invited Lecturer at the Non-convex Optimization for Machine Learning (NOML) Summer School, IIT Bombay, India, June 2015

  • Successfully defended in March 2015!

Selected Publications

For the complete list of publications and workshop papers, see my publications page or my research page.

Funding and Support

Our research is supported graciously by research grants from NSF, Google, Adobe, Amazon, and the UT Dallas startup fund. See our funding page for more details.