Hi! I'm Pengcheng. I am a research scientist at Google Deepmind, working on Gemini Code and research projects in natural language to code generation. Before that, I was a Ph.D. student at Language Technologies Institute, Carnegie Mellon University.
Outside of work, I am a student pilot based at Palo Alto Airport. I also train Coke using RLHF :)
Research Papers
Please see my Google Scholar page for recent publications.Past Industrial Experiences
- Researcher Intern, Microsoft Semantic Machines
- Part-time Research Collaborator, Facebook AI Research
- Research Intern, Facebook AI Research London
- Research Intern, Microsoft Research Cambridge, UK
- Research Intern, Microsoft Research
- Research Intern, Noah's Ark Lab, Huawei
- Research Intern, Microsoft Research Asia
Professional Services
- Area Chair: ICLR 2024, ICLR 2025, ACL 2025
- Reviewer: ACL (outstanding reviewer @ ACL 2020), EMNLP, NAACL, NeurIPS, ICML (top 33% reviewer @ ICML '20), ICLR, etc.
Talks and Coding
- Stanford CS224N DL with NLP: Code Generation (slides)
- TranX: a general-purpose syntax-driven neural semantic parser
- Strong results on six semantic parsing benchmarks
- pytorch_basic_nmt: a basic implementation of attentional nerual seq2seq models
- Used for instructional purposes in Stanford CS224N Nautral Language Processing with Deep Learning and CMU 11-731 Machine Translation and Sequence-to-Sequence Models.