Ying Sheng(盛颖)

I am a final-year Ph.D. student at the Computer Science Department at Stanford University, affiliated with Centaur. I am very fortunate to be advised by Clark Barrett. Prior to that, I received an M.S. in Computer Science from Columbia University in 2017 and a B.E. in Computer Science and Technology from ACM Honored Class, Shanghai Jiao Tong University in 2016.

I'm visiting Sky@UC Berkeley, working with Ion Stoica and Joseph E. Gonzalez. Before that, I was a Ph.D. resident at X, the moonshot Factory (the team graduated to Labs@Google during my residence) (2022), working on AI for Code with Michele Catasta; a research intern at Facebook Novi (2021), working on smart contract verification with Prof. David Dill; a quantitative software engineer at Two Sigma (2018); and a research intern at Microsoft Research Asia (2015), working with Chin-Yew Lin.

NEWS:

Contact: ying1123@stanford.edu,  or sqy1415@gmail.com.

Github: https://github.com/Ying1123

Research Interests:

        Large Language Models, Machine Learning Systems, Machine Learning for Programs.

Formal Methods, Automated Reasoning.

        Code Generation.

        I like collaboration and competition. :)

Projects:

SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.

S-LoRA is a lightweight system for serving thousands of LoRA adapters concurrently.

cvc5 is an open-source automatic theorem prover for Satisfiability Modulo Theories (SMT) problems.

FlexGen is a high-throughput generation engine for running large language models with limited GPU memory. 

FastChat is an open platform for training, serving, and evaluating large language model based chatbots.

Vicuna is a chatbot trained by fine-tuning LLaMA on user-shared conversations.

Selected Awards

Teaching