About me

Hello! I am now an Applied Scientist in Amazon Rufus (Palo Alto, US).

I got a Ph.D. from School of Information, University of Michigan, advised by Prof. VG Vinod Vydiswaran. I also worked with Prof. Qiaozhu Mei between 2015-2020, and Prof. Daniel M. Romero between 2018-2020.

Prior to UM, I received my honored bachelor degree in Computer Science from Chu Kochen Honors College, Zhejiang University in 2017. During my undergraduate period, I was supervised by Prof. Fei Wu.

I primarily working on Natural Language Processing and Machine Learning. My recent research focuses on below topics:

  • Self-Supervised Representation Learning (contrastive learning).

  • Historical Language Change (such as semantic shift)

  • Robustness of ML models (backdoor attack and defense of LLMs).

  • Diffusion Models for Non-Autoregressive Generation.

  • Knowledge Distillation (from advanced black-box LLM to small models).

News

[Sep 2024] Our paper Divide-or-Conquer? Which Part Should You Distill Your LLM? is accepted by EMNLP 2024.

[May 2024] Joined the Amazon Rufus team.

[May 2024] Got married and hosted our wedding!

[Apr 2024] Defended my thesis and left Ann Arbor, a city where I lived for over 8 years.

[Feb 2024] Our paper Divide-or-Conquer? Which Part Should You Distill Your LLM? is now available for public access.

[Oct 2023] Our paper HiCL: Hierarchical Contrastive Learning of Unsupervised Sentence Embeddings is accepted by EMNLP 2023.

[Sep 2023] Our paper PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model is accepted by NeurIPS 2023.

[Aug 2023] Finished my internship at Machine Learning Research at Apple.