About Me

I’m a Senior Researcher at Microsoft Research. I received my Bachelor’s degree and Ph.D. from Tsinghua University in 2016 and 2021, respectively. My recent research focuses on computer use agents, tool calling, and agentic AI.

📢 I’m looking for a self-motivated research intern who is passionate about AI agents (GUI/embodied agents, tool calling, and context management). If you are interested in these topics, feel free to email me your resume and a brief self-intro!

News 🌟

Jan. 2026 Welcome to join our 5th Workshop on Computer Vision in the Wild (CVinW) at @CVPR 2026!
Jan. 2026 We release SynthAgent, a task and trajectory synthetic framework for web agents!
Jan. 2026 Dyna-Mind is accepted by ICLR 2026!
Dec. 2025 We release Argos, a principled reward agent to train LMRMs for agentic tasks.
Nov. 2025 We release GUI-360, a comprehensive dataset and benchmark for CUA!
Jun. 2025 Excited to release GUI-Actor, a coordinate-free visual grounding method for GUI agents!

Show more news

Apr. 2025 You’re welcome to join our 4th Workshop on Computer Vision in the Wild (CVinW) at @CVPR 2025!
Apr. 2025 We release MMInference, accelerating pre-filling for long-context VLMs!
Feb. 2025 We release Magma, a foundation model for multimodal AI agents!
Jan. 2025 SeCom and SCBench are accepted at ICLR 2025!
Sep. 2024 MInference is accepted at NeurIPS 2024 as a spotlight!
Sep. 2024 I’m serving as an Area Chair for COLING 25!
Jun. 2024 MInference and LLM Position Bias paper are accepted to ES-FoMo II @ ICML24 and LCFM @ ICML24, respectively.
May. 2024 LLMLingua Series has been integrated as a custom tool in Prompt Flow, AutoGen, LangChain and LlamaIndex.
May. 2024 LongLLMLingua and LLMLingua-2 are accepted to ACL-2024 in main track and findings!
Mar. 2024 We release LLMLingua-2, an efficient option for task-agnostic prompt compression with good performance and generalizability across different scenarios, boasting a 3x-6x speed improvement over LLMLingua!
Oct. 2023 We release LongLLMLingua, aiming to accelerate and enhance LLM inference in long-context scenarios via question-aware prompt compression and content reorganization!
Oct. 2023 We release LLMLingua, a coarse-to-fine prompt compression method based on perplexity from a small language model such as LLaMA-7B!

Selected Publications 📚

Agentic AI
Efficient LLM Inference
Information Extraction & Low-Resource NLP

Honors & Awards 🏆

2024 Microsoft 2024 Global Hackathon Executive Challenge Winner
2024 Microsoft Machine Learning, AI & Data Science Conference (MLADS) Distinguished Contribution
2023 Microsoft 2023 Global Hackathon Award Winner
2020 Outstanding Intern of “Stars of Tomorrow” Program, Microsoft Research Asia
2020 Intel Scholarship
2018 Outstanding (12-9) Counselor Prize, Tsinghua University
2016 Outstanding Bachelor Thesis, Tsinghua University
2014 National Encouragement Scholarship
2014 Scholarship of Art Excellence, Tsinghua University
2013 Scholarship of Academic Excellence, Tsinghua University

Other Information 📝

Invited Talks

Sep. 2025 Towards AI Agents That Can See And Act @ Shanghai Artificial Intelligence Laboratory
Jun. 2025 Act Where You See: Coordinate-Free Visual Grounding for GUI Agents @ Simular Seminar

Academic Service

  • Area Chair: COLING-25.
  • Conference Reviewer: ICLR, NeurIPS, CVPR, ACL, EMNLP, NAACL, AAAI, NLPCC.
  • Journal Reviewer: CSUR, TOIS, Pattern Recognition, TASLP, IPM, JIM, ESIN, SIVP.

Other Activities

Sep. 2016 - Aug. 2018 Counselor at Center for Student Learning and Development, Tsinghua University
Aug. 2012 - Jun. 2021 Member of Chinese National Orchestra, Tsinghua University