I am a postdoctoral researcher at CMU working with Prof. Graham Neubig on natural language processing (NLP) and large language models (LLMs).
I obtained my Ph.D. from The Ohio State University (OSU-NLP-Group), where I was trained by Prof. Huan Sun and Prof. Yu Su.
I also collaborate closely with Prof. Wenhu Chen at University of Waterloo.
My research aims to understand and enhance the reasoning capabilities of LLMs while improving their responsibility and reliability.
- Understanding and enhancing the reasoning capabilities of (multimodal) LLMs
- Code generation for autonomous programming and reasoning
- Improving the responsibility of LLMs in areas such as privacy, robustness, and attribution.
- LLMs for science and healthcare
You might have heard of me because of the following work:
- MMMU / MMMU-Pro: the commonly used multimodal language model evaluation suite. 🔥[Slides]
- MAmmoTH / MAmmoTH2: Strong reasoning models achieving SoTA in 2023 and 2024.
- OpenCodeInterpreter: the open replication of OpenAI GPT-4 Code Interpreter to achieve SoTA on code generation tasks.
- MixEval: a fast, cheap, and highly Arena correlated LLM benchmark.
- MMLU-Pro: A more robust and challenging version of MMLU featuring more on reasoning.
- Grokked Transformers: A mechanistic analysis revealing that why transformers could implicit reason through grokking.