
<aside> ๐ Google Scholar
</aside>
<aside> <img src="https://prod-files-secure.s3.us-west-2.amazonaws.com/27eb5805-53fb-4a73-8248-9cca33adce3c/f30304aa-3449-4968-94ed-4711f30b0b92/icons8-twitter-48.png" alt="https://prod-files-secure.s3.us-west-2.amazonaws.com/27eb5805-53fb-4a73-8248-9cca33adce3c/f30304aa-3449-4968-94ed-4711f30b0b92/icons8-twitter-48.png" width="40px" /> Twitter
</aside>
I also maintain most of my study notes on LLM here.
$$ \Large \textbf {About Me} \\ $$
I did my Ph.D. studies at Department of Computer Science, UCLA, where I am very fortunate to be advised byย Prof. Wei Wang.
Earlier, I completed research internships at **Google DeepMind,** **Google,** Microsoft Research and **Amazon AWS.** I earned both my B.S. in Mathematics and M.S. in Computer Science from UCLA. During that time, Iโve been an student researcher atย UCLA-NLPย group withย Prof. Kai-Wei Chang.
I work on multi-modal LLMs.
Quick links:
2025

Member of Technical Staff | xAI
Multi-modal
2025

Student Researcher | Google DeepMind
Synthetic Data for LLM Post-training
2025

Student Researcher | Google LLC
RL for LLM Agentic Reasoning
Project 1: Ultra-fast Exploration for Scalable Agentic Reasoner [Internal Contribution to Google LLMs]
Project 2: Supervised Reinforcement Learning: From Expert Trajectories to Step-wise Reasoning [arXiv]
2024

Research Intern | Microsoft Research
LLM Self-Training for Math Reasoning.
Paper: Flow-DPO: Improving LLM Mathematical Reasoning through Online Multi-Agent Learning
[NeurIPS 2024 Math-AI Workshop] [Media]
2023

Applied Scientist Intern | Amazon AWS
Large Language Model Reasoning with Knowledge Graphs.
Multi-modal LLMs
<aside> ๐

</aside>
<aside> ๐

</aside>
<aside> ๐

</aside>
<aside> ๐
Synthetic Data for LLM Improvement
<aside> ๐

</aside>
<aside> ๐
<aside> ๐

</aside>
<aside> ๐

</aside>