๐ Hi, I'm Ming Zhang (ๅผ ๆ), also known as DinoBro (ๆ้พๅฅ) or DinoDoctor (ๆ้พๅๅฃซ). I am a third-year direct Ph.D. student at the FudanNLP Lab, School of Computer Science, Fudan University, co-advised by Prof. Qi Zhang, A.P. Tao Gui, and Prof. Xuanjing Huang.
๐ฌ My research focuses on Large Language Model Evaluation and Dialogue Systems. Recently, I have been particularly interested in Context Learning and AI for Academia.
๐ข I previously interned at ByteDance (2024.09 โ 2025.09) and am currently interning at the Tencent Qingyun Program (2025.12 โ Present).
๐ I serve as a Reviewer for AAAI, ACL ARR, ICLR, NeurIPS, and ICML, and as an Area Chair for ACL ARR.
Beyond research, my personal interests include:
๐ญ I have been a lifelong astronomy enthusiast and served as President of the Fudan Astronomy Society during my undergraduate years.
โฝ I am a devoted fan of football โ my all-time idol is Lionel Andrรฉs Messi.
๐ฎ I served as Captain of the Fudan University League of Legends Varsity Team (Jungle), peaked at Challenger (ๆๅผบ็่ ) in S3 and S12, and my favorite champion lately is Gangplank.
๐ง mingzhang23 [at] m [dot] fudan [dot] edu [dot] cn / konglongge [at] outlook [dot] com
๐ฌ Please feel free to add me on WeChat: zanyingluan
๐ฅ News
- 2026.02 ๐ Thinking with Video accepted by CVPR 2026!
- 2026.02 ๐ CL-bench is now available on arXiv!
- 2026.01 ๐ OpenNovelty and TaxoBench are now available on arXiv!
- 2026.01 ๐ Game-RL accepted by ICLR 2026!
- 2025.11 ๐ Reasoning or Memorization and Speech Tokenizer accepted by AAAI 2026!
- 2025.09 ๐ EvaLearn accepted by NeurIPS 2025!
- 2025.08 ๐ LLMEval-Med accepted by EMNLP 2025!
- 2025.05 ๐ PFDial accepted by ACL 2025!
- 2025.01 ๐ Our LLM Agent Survey published in Science China Information Sciences!
- 2024.09 ๐ TransferTOD and MathTrap accepted by EMNLP 2024!
- 2024.07 ๐ Mousi accepted by COLM 2024!
- 2023.12 ๐ LLMEval accepted by AAAI 2024!
โญ Selected Works
* denotes co-first author, โ denotes corresponding author.
CL-bench: A Benchmark for Context Learning
A comprehensive benchmark for evaluating context learning capabilities of large language models, providing systematic assessment across diverse context-dependent tasks.
EvaLearn: Quantifying the Learning Capability and Efficiency of LLMs via Sequential Problem Solving NeurIPS 2025
A novel framework for quantifying the learning capability and efficiency of large language models through sequential problem solving, providing new insights into how LLMs acquire and apply knowledge.
The Rise and Potential of Large Language Model Based Agents: A Survey SCIS
A comprehensive survey on LLM-based agents covering their construction, applications, and evaluation. This highly influential work provides a systematic overview of the emerging field of autonomous agents powered by large language models. Published in Science China Information Sciences.