Welcome


I am a joint J.D.-Ph.D. student at Stanford University, advised by Professor Dan Jurafsky in the CS Department and Professor Daniel E. Ho at Stanford Law School. 

I am also pursuing a Ph.D. Minor in the Philosophy, Language, and the Arts program; and working as a Graduate Research Fellow at the Regulation, Evaluation, and Governance Lab (RegLab). Previously, I completed my undergraduate studies at Harvard College, pursuing a joint degree in Mathematics and Computer Science and a secondary in Folklore & Mythology.

[Google Scholar] / [Semantic Scholar] / [GitHub] / [string2string Library] / [BIG-Bench Hard]

ML & NLP Research

I am broadly interested in natural-language processing (NLP) and machine learning (ML).

My current research endeavours in NLP cover the following topics:

Selected Publications

Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding

Mirac Suzgun, Adam Tauman Kalai

(under review)

[paper] / [GitHub]

Large Legal Fictions: Profiling Legal Hallucinations in Large Language Models

Matthew Dahl, Varun Magesh, Mirac Suzgun, Daniel E. Ho

Journal of Legal Analysis 

[paper] / [GitHub] / [Bloomberg Law News] / [Stanford HAI blogpost]

Assessing the Potential of GPT-4 to Perpetuate Racial and Gender Biases in Health Care:         A Model Evaluation Study

Travis Zack, Eric Lehman, Mirac Suzgun, Jorge A. Rodriguez, Leo Anthony Celi, Judy Gichoya, Dan Jurafsky, Peter Szolovits, David W. Bates, Raja-Elie E. Abdulnour, Atul J. Butte,  Emily Alsentzer

The Lancet 

[paper] / [pubmed] / [STAT News] / [GitHub]

Do Language Models Know When They're Hallucinating References?

Ayush Agrawal, Mirac Suzgun, Lester Mackey, Adam Tauman Kalai

EACL 2024

[paper] / [GitHub]

Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions

Federico Bianchi, Mirac Suzgun, Giuseppe Attanasio, Paul Röttger, Dan Jurafsky, Tatsunori Hashimoto, James Zou

ICLR 2024

[paper] / [GitHub]

A Benchmark for Learning to Translate a New Language from One Grammar Book

Garrett Tanzer, Mirac Suzgun, Eline Visser, Dan Jurafsky, Luke Melas-Kyriazi

ICLR 2024 (Oral Presentation)

[paper] / [GitHub] / [Gemini Blogpost] / [Jeff Dean's Twitter post]

string2string: A Modern Python Library for String-to-String Algorithms

Mirac Suzgun, Stuart M. Shieber, Dan Jurafsky

arXiv 2023 

[paper] / [GitHub] / [documentation] / [pip install string2string]

Follow the Wisdom of the Crowd: Effective Text Generation via Minimum Bayes Risk Decoding

Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky

ACL 2023 (Findings)

[paper] / [code] / [colab]

Scaling Instruction-Finetuned Language Models

Hyung Won Chung, Le Hou, Shayne Longpre, Barret Zoph, Yi Tay, William Fedus, Eric Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson, Shixiang Shane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, Aakanksha Chowdhery, Sharan Narang, Gaurav Mishra, Adams Yu, Vincent Zhao, Yanping Huang, Andrew Dai, Hongkun Yu, Slav Petrov, Ed H. Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, Quoc V. Le, Jason Wei

JMLR 2024

[paper] / [checkpoints] / [Flan-T5 models]

Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them

Mirac Suzgun, Nathan Scales, Nathanael Schärli, Sebastian Gehrmann, Yi Tay, Hyung Won Chung, Aakanksha Chowdhery, Quoc V. Le, Ed H. Chi, Denny Zhou, Jason Wei

ACL 2023 (Findings)

[paper] / [code] 

Language Models are Multilingual Chain-of-Thought Reasoners

Freda Shi, Mirac Suzgun, Markus Freitag, Xuezhi Wang, Suraj Srivats, Soroush Vosoughi, Hyung Won Chung, Yi Tay, Sebastian Ruder, Denny Zhou, Dipanjan Das, Jason Wei

ICLR 2023

[paper] / [dataset and code] 

When Do Pre-Training Biases Propagate to Downstream Tasks? A Case Study in Text Summarization

Faisal Ladhak, Esin Durmus, Mirac Suzgun, Tianyi Zhang, Dan Jurafsky, Kathleen McKeown, Tatsunori Hashimoto

EACL 2023

[paper]  / [code] 

The Harvard USPTO Patent Dataset:                         A Large-Scale, Well-Structured, and Multi-Purpose Corpus of Patent Applications

Mirac Suzgun, Luke Melas-Kyriazi, Suproteem K. Sarkar, Scott Duke Kominers, Stuart M. Shieber

NeurIPS Datasets and Benchmarks 2023 (Spotlight)

[paper] / [code] / [colab] / [website] / [poster]

Prompt-and-Rerank:         A Method for Zero-Shot and Few-Shot Arbitrary Textual Style Transfer with Small Language Models

Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky

EMNLP 2022 (Oral Presentation)

[paper] / [code] 

Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models

Google Research's BIG-Bench Effort 

[Contributed to the Dyck Languages and Dynamic Counting tasks]

TMLR 2023

[paper] / [code and data]

Monte Carlo Tree Search for Interpreting Stress in Natural Language

Kyle Swanson, Joy Hsu, Mirac Suzgun

ACL 2022 Workshop on LT-EDI (Oral Presentation)

[paper] / [code]

Memory-Augmented Recurrent Neural Networks Can Learn Generalized Dyck Languages

Mirac Suzgun, Sebastian Gehrmann, Yonatan Belinkov, Stuart M. Shieber

arXiv 2019

[paper] / [code]

LSTM Networks Can Perform Dynamic Counting

Mirac Suzgun, Sebastian Gehrmann, Yonatan Belinkov, Stuart M. Shieber

ACL 2019 Workshop on Deep Learning and Formal Languages (Spotlight)

[paper] / [code]

On Evaluating the Generalization of LSTM Models in Formal Languages

Mirac Suzgun, Yonatan Belinkov, Stuart M. Shieber

SCiL 2019

[paper] / [code]

Formal Language Theory as a Framework for Understanding the Limitations of Recurrent Neural Networks

Mirac Suzgun

Undergrate Thesis (advised by Stuart M. Shieber and Peter B. Kronheimer). Awarded the Thomas T. Hoopes Prize.

Available upon request.