CV
Education
- Ph.D in Information Science, Drexel University, 2026 (expected)
- M.S. in Information Science, University of North Carolina at Chapel Hill, 2021
- B.S. in Management Information System, Dalian University of Technology, 2019
Work experience
NLP Research Intern
- Samsung Research America, Mountain View, California (September 2025 – now)
- Working on improving QA Grounding with layout information.
Applied Scientist Intern
- Amazon, Seattle, Washington (June 2025 – September 2025)
- Designed and implemented a novel framework that generates precise product category definitions through reinforcement learning from verbal feedback, replacing costly manual definition writing
- Built a modular system leveraging Claude Sonnet 3.5 and Claude Haiku 3.5, including components for the actor, evaluators, self-reflection, and a memory module to store historical reflections
- Designed a sampling model to extract representative product examples from millions of listings, enabling scalable learning
- Achieved 97.5% classification accuracy (vs. 94% human-crafted definitions) while generating shorter, clearer outputs
- Reduced category definition creation from weeks to hours, accelerating business response to new product categories and emerging market demands
Research Scientist Intern
- Ping An Technology Research Lab, Palo Alto, California (March 2025 – June 2025)
- Designed and optimized both ToB and ToC medical VLM, supporting real-world deployment in general practice settings
- Fine-tuned 32B, 72B Qwen vision-language large models (VLMs) on 400K multi-turn medical consultation dialogues using LoRA and DeepSpeed on 8×A800 GPUs
- Developed a RAG (Retrieval-Augmented Generation) pipeline with a knowledge graph of rare diseases and treatments, improving diagnostic accuracy from 82% to 90% while reducing hallucinations
- Quantized Qwen 72B VLM with GPTQ, decreasing model size by 69.65%
- Deployed the quantized model using vLLM for inference acceleration, reducing response latency by 50–66% compared to baseline
Skills
- Model Training
- TensorFlow, PyTorch
- SFT, RLHF, PPO, QLoRA
- Distributed RPC (Data/Model Parallel)
- Quantization Aware Training
- Model Deployment
- Docker, Kubernetes, Tensorflow Lite
- React.js, Django, Node.js, Spring Cloud
- AWS EC2
- Programming Languages
- Python (transformers, opencv, nltk, sklearn, scipy)
- Java, JavaScript, C, C#
- Shell/Scripting
- SQL/NoSQL
Service
- Reviewer for the AAAI Conference on Artificial Intelligence (AAAI), 2025
- Reviewer for the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2025
- Reviewer for the 63rd Annual Meetings of the Association for Computing Linguistics (ACL), 2025
- Reviewer for International Conference on Computational Linguistics (COLING), 2024
Academic Activities
- Oral Presentation at COLING, January 2025
- Oral Presentation at IJCAI XAI workshop, June 2024
- Poster Presentation at AAAI Doctoral Consortium, February 2025
- Poster Presentation at AAAS, February 2024
- Poster Presentation at AAAI IAAI session, Washington D.C., 2023
Awards & Scholarships
- Phoebe W. Haas Endowed Fellowship for Women Doctoral Students, 2023
- Honorable Mention (2nd place) in Mathematical Contest in Modeling (MCM), USA, 2017